Chat (Fireworks AI - llama-v2-70b-chat) 🏆

POST https://api.fireworks.ai/inference/v1/chat/completions

Request Body

{"messages"=>[{"role"=>"user", "content"=>"Explain the importance of low latency LLMs"}], "model"=>"accounts/fireworks/models/llama-v2-70b-chat", "temperature"=>0.5, "max_tokens"=>1024, "top_p"=>1, "stream"=>true, "stop"=>nil}

RESPONSES

status: OK

"data: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"role\":\"assistant\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"Hello!\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" I'm here to help you understand the importance of low latency\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" LLMs.\\n\\nLLMs, or large language\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" models, are artificial intelligence models that are trained on vast amounts of text data\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" to generate language outputs that\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" are coherent and natural-sounding.\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" Low latency LLMs are those that are designed to respond quickly\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" and efficiently, often in real-time, to user input\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\".\\n\\nThe importance of low latency LLMs lies in their ability\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" to enable real-time communication and interaction between humans and machines. With low\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" latency LLMs, it becomes possible to build convers\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"ational AI systems that can respond to user queries and\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" requests in a timely manner, making them\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" feel more natural and human-like.\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"\\n\\nLow latency LLMs have numerous\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" applications in various industries, such as:\\n\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"\\n1. Customer Service: Chatb\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"ots powered by low latency LLMs can provide instant responses to customer\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" inquiries, improving customer satisfaction and reducing the need for human customer\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" support.\\n2. Virtual Assistants\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\": Virtual assistants like Siri, Alexa, and\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" Google Assistant use low latency LLMs to respond to voice commands\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\", making them feel more responsive and intuitive.\\n\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"3. Language Translation: Low latency LLMs can be used\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" to translate text and speech in real-time, breaking language\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" barriers and enabling communication between people who speak different languages.\\n4\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\". Gaming: Low latency LLMs can\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" be used to create more immersive gaming experiences, such\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" as real-time dialogue and voice-\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"controlled gameplay.\\n5. Education:\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" Low latency LLMs can be used to create personalized\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" learning experiences, such as virtual teaching assistants\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" that can respond to student questions and provide real-\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"time feedback.\\n\\nIn summary, low latency LLMs are essential for building convers\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"ational AI systems that can interact with humans in a natural and responsive\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\" manner. They have numerous applications across various industries and have the potential to revolution\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{\"content\":\"ize the way we interact with machines.\"},\"finish_reason\":null}],\"usage\":null}\n\ndata: {\"id\":\"ddc30dd6-b674-4cbf-9fad-0957b4644214\",\"object\":\"chat.completion.chunk\",\"created\":1712018387,\"model\":\"accounts/fireworks/models/llama-v2-70b-chat\",\"choices\":[{\"index\":0,\"delta\":{},\"finish_reason\":\"stop\"}],\"usage\":{\"prompt_tokens\":100,\"total_tokens\":541,\"completion_tokens\":441}}\n\ndata: [DONE]\n\n"