added youtube videos
parent
3168dbf4b8
commit
215bb7377b
Binary file not shown.
After Width: | Height: | Size: 156 KiB |
@ -0,0 +1,21 @@
|
||||
# Reducing Hallucination in Structured Outputs via RAG
|
||||
|
||||
import {Bleed} from 'nextra-theme-docs'
|
||||
|
||||
<Bleed>
|
||||
<iframe width="100%"
|
||||
height="415px"
|
||||
src="https://www.youtube.com/embed/TUL5guqZejw?si=Doc7lzyAY-SKr21L" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture"
|
||||
allowFullScreen
|
||||
/>
|
||||
</Bleed>
|
||||
|
||||
Researchers at ServiceNow shared a [new paper](https://arxiv.org/abs/2404.08189) where they discuss how to deploy an efficient RAG system for structured output tasks.
|
||||
|
||||
!["RAG Hallucination"](../../img/research/structured_outputs.png)
|
||||
|
||||
The RAG system combines a small language model with a very small retriever. It shows that RAG can enable deploying powerful LLM-powered systems in limited-resource settings while mitigating issues like hallucination and increasing the reliability of outputs.
|
||||
|
||||
The paper covers the very useful enterprise application of translating natural language requirements to workflows (formatted in JSON). So much productivity can come from this task but there is a lot of optimization that can be further achieved (eg., using speculative decoding or using YAML instead of JSON).
|
||||
|
||||
The paper provides some great insights and practical tips on how to effectively develop RAG systems for the real world.
|
@ -0,0 +1,21 @@
|
||||
# Best Practices and Lessons Learned on Synthetic Data for Language Models
|
||||
|
||||
import {Bleed} from 'nextra-theme-docs'
|
||||
|
||||
<Bleed>
|
||||
<iframe width="100%"
|
||||
height="415px"
|
||||
src="https://www.youtube.com/embed/YnlArBZJHY8?si=ZH3hFzwixUopxU5Z" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture"
|
||||
allowFullScreen
|
||||
/>
|
||||
</Bleed>
|
||||
|
||||
This [paper](https://arxiv.org/abs/2404.07503) provides an overview of best practices and lessons learned on synthetic data for language models ans was published by Google DeepMind and other collaborators.
|
||||
|
||||
It focuses on synthetic data and covers applications, challenges, and future directions. This is an important paper given the significant advancements we are seeing from the use of synthetic data in the field of AI.
|
||||
|
||||
We know for sure that the more high-quality data we give these models, the better the performance. Creating synthetic data is not hard but ensuring its quality is really the challenge.
|
||||
|
||||
The paper also discusses important topics when working with synthetic data such as ensuring quality, factuality, fidelity, unbiasedness, trustworthiness, privacy, and more.
|
||||
|
||||
There are a lot of great references mentioned in the related work section as well.
|
Loading…
Reference in New Issue