Meet Alibaba 'Marco-o1' That Wants To Redefine AI Reasoning By Making It Open Source

Alibaba Marco-o1

The AI industry was once a quiet, niche space, rarely making waves beyond its own sphere.

That all changed when OpenAI launched ChatGPT. Since then, tech companies big and small have been racing to create cutting-edge AI tools to address a wide range of needs and applications.

While many innovations have emerged from the West, the East is proving it’s not far behind in this competitive landscape.

Alibaba is one of the most prominent tech companies from China, and here, its researchers revealed what is supposed to be a direct competitor to OpenAI-o1, a specialized Large Language Model capable of reasoning.

Alibaba calls its own take the 'Marco-o1'.

Alibaba Marco-o1

According to its Hugging Face page and GitHub page:

"Marco-o1 not only focuses on disciplines with standard answers, such as mathematics, physics, and coding—which are well-suited for reinforcement learning (RL)—but also places greater emphasis on open-ended resolutions."

The idea is to be an AI capable of solving complex problems classic language models often struggle with.

More or less, Marco-o1 is designed to be just like OpenAI-o1

In fact, the researchers did say that they were inspired by OpenAI-o1, saying that:

"OpenAI recently introduced the groundbreaking o1 model, renowned for its exceptional reasoning capabilities. This model has demonstrated outstanding performance on platforms such as AIME, CodeForces, surpassing other leading models. Inspired by this success, we aimed to push the boundaries of LLMs even further, enhancing their reasoning abilities to tackle complex, real-world challenges."

But Marco-o1 is more than just a OpenAI-o1 copy because the researchers said that it leverages advanced techniques that can distinguish itself.

Read: Alibaba Introduces MIMO, A Way To Create A Controllable Character Through AI Video Synthesis

Alibaba Marco-o1

Highlights include:

  1. Fine-Tuning with CoT Data: The researchers created Marco-o1-CoT by performing full-parameter fine-tuning on the base model using open-source CoT dataset combined with their own self-developed synthetic data.
  2. Solution Space Expansion via MCTS: The researchers integrated LLMs with MCTS (Marco-o1-MCTS), using the model's output confidence to guide the search and expand the solution space.
  3. Reasoning Action Strategy: The AI implements novel reasoning action strategies and a reflection mechanism (Marco-o1-MCTS Mini-Step), including exploring different action granularities within the MCTS framework and prompting the model to self-reflect, thereby significantly enhancing the model's ability to solve complex problems.
  4. Application in Translation Tasks: The researchers claim that Marco-o1 is the first to apply Large Reasoning Models (LRM) to Machine Translation task, exploring inference time scaling laws in the multilingual and translation domain.

This allows it to have an exceptionally advanced ability in handling complex tasks that traditional LLMs may have difficulties with.

"MCTS allows exploration of multiple reasoning paths using confidence scores derived from softmax-applied log probabilities of the top-k alternative tokens, guiding the model to optimal solutions. Moreover, our reasoning action strategy involves varying the granularity of actions within steps and mini-steps to optimize search efficiency and accuracy," the researchers said.

Published: 
26/11/2024