4242from materializers .document_materializer import DocumentMaterializer
4343from pipelines import (
4444 finetune_embeddings ,
45+ generate_synthetic_data ,
4546 llm_basic_rag ,
4647 llm_eval ,
4748)
102103 default = False ,
103104 help = "Disable cache." ,
104105)
106+ @click .option (
107+ "--synthetic" ,
108+ "synthetic" ,
109+ is_flag = True ,
110+ default = False ,
111+ help = "Run the synthetic data pipeline." ,
112+ )
105113@click .option (
106114 "--local" ,
107115 "local" ,
@@ -143,6 +151,7 @@ def main(
143151 query : Optional [str ] = None ,
144152 model : str = OPENAI_MODEL ,
145153 no_cache : bool = False ,
154+ synthetic : bool = False ,
146155 local : bool = False ,
147156 embeddings : bool = False ,
148157 dummyembeddings : bool = False ,
@@ -157,6 +166,7 @@ def main(
157166 query (Optional[str]): If provided, the RAG model will be queried with this string.
158167 model (str): The model to use for the completion. Default is OPENAI_MODEL.
159168 no_cache (bool): If `True`, cache will be disabled.
169+ synthetic (bool): If `True`, the synthetic data pipeline will be run.
160170 local (bool): If `True`, the local LLM via Ollama will be used.
161171 dummyembeddings (bool): If `True`, dummyembeddings will be used
162172 embeddings (bool): If `True`, the embeddings will be fine-tuned.
@@ -188,6 +198,8 @@ def main(
188198 llm_basic_rag .with_options (** pipeline_args )()
189199 if evaluation :
190200 llm_eval .with_options (** pipeline_args )()
201+ if synthetic :
202+ generate_synthetic_data .with_options (** pipeline_args )()
191203 if embeddings :
192204 finetune_embeddings .with_options (** embeddings_finetune_args )()
193205 if dummyembeddings :
0 commit comments