Blue v0.9: “Agentic” for the Enterprise!

Blue: "Agentic" for the Enterprise Blue v0.9 vol 1

We present Blue v0.9, our open-source framework for building and deploying agentic workflows in enterprise environments. Unlike conventional AI frameworks, Blue is designed with enterprise-scale requirements in mind—scalability, observability, configurability, and seamless integration with existing infrastructure.

EMNLP 2024 Highlights

Megagon Labs has curated this blog post to spotlight key developments in agentic systems, AI safety, and human-centered AI.

Optimizing Compound AI Systems

Optimizing Compound AI Systems

We echo through this blog that the optimization framework for compound AI systems should achieve broader goals such as multi-objective (accuracy, cost, latency, etc.), multi-plan optimization and also handling constraints, especially the budget. Again, these optimization goals are not comprehensive by far but are important for enterprise scenarios.

AmbigNLG: A Tutorial

Ambig NLG Tutorial

AmbigNLG tackles ambiguity in Natural Language Generation (NLG) instructions by identifying unclear specifications and refining them for better output quality.

Megagon Labs Summer 2024 Internship Experience

2024 Summer Internship Experience at Megagon Labs

Through this inside peek at our internship program, explore the types of projects we at Megagon Labs formulate for our interns. If you are looking to start an internship soon, take their advice and apply it to your own internships.

NAACL 2024 Highlights & Big Trends Shaping NLP

NAACL 2024 Highlights & Big Trends Shaping NLP

Drawing from our experience at the NAACL conference, the Megagon Labs team has crafted this blog post to highlight three major trends: targeted evaluation, reasoning, and fine-tuning/RAG. These trends represent significant advancements in the field of NLP and showcase the innovative approaches researchers are taking to enhance the capabilities of LLMs.

Order Matters: Assessing LLM Sensitivity in Multiple-Choice Tasks

Order Matters: Assessing LLM Sensitivity in Multiple-Choice Tasks

Explore the relationship between option arrangement and performance variations in Large Language Models (LLMs) during multiple-choice tasks. Through meticulous analysis, we uncovered substantial sensitivity of LLMs to the order of answer options, with performance fluctuations of up to 75% across different benchmarks.