@@ -59,18 +59,15 @@ async function main() {
5959 console . log ( '\n*** Generating text:' ) ;
6060
6161 // Convert config to Vercel AI SDK format
62- const vercelConfig = VercelProvider . toVercelAISDK (
63- aiConfig ,
64- openai ,
65- { nonInterpolatedMessages : [ userMessage ] } ,
66- ) ;
62+ const vercelConfig = VercelProvider . toVercelAISDK ( aiConfig , openai , {
63+ nonInterpolatedMessages : [ userMessage ] ,
64+ } ) ;
6765
6866 // Track metrics using trackMetricsOf with VercelProvider.createAIMetrics
69- const result = await aiConfig . tracker . trackMetricsOf (
70- VercelProvider . createAIMetrics ,
71- ( ) => generateText ( vercelConfig ) ,
67+ const result = await aiConfig . tracker . trackMetricsOf ( VercelProvider . createAIMetrics , ( ) =>
68+ generateText ( vercelConfig ) ,
7269 ) ;
73-
70+
7471 console . log ( 'Response:' , result . text ) ;
7572 } catch ( err ) {
7673 console . error ( 'Error:' , err ) ;
@@ -86,12 +83,10 @@ async function main() {
8683 // Example of using generateText (non-streaming)
8784 console . log ( '\n*** Streaming text:' ) ;
8885 // Convert config to Vercel AI SDK format
89- const vercelConfig = VercelProvider . toVercelAISDK (
90- aiConfig ,
91- openai ,
92- { nonInterpolatedMessages : [ userMessage ] } ,
93- ) ;
94-
86+ const vercelConfig = VercelProvider . toVercelAISDK ( aiConfig , openai , {
87+ nonInterpolatedMessages : [ userMessage ] ,
88+ } ) ;
89+
9590 // Track streaming metrics using trackStreamMetricsOf with provider's extractor
9691 // Stream is returned immediately (synchronously), metrics tracked in background
9792 const streamResult = aiConfig . tracker . trackStreamMetricsOf (
0 commit comments