Curated Resource ( ? )

A Quiet Revolution? Mistral AI Releases Sensational New AI Model

my notes ( ? )

"startup Mistral AI posted a download link to their latest language model". In many ways it's equivalent to GPT4: context size of 32k tokens, and built using the “Mixture Of Experts” model, combining "several highly specialized language models... 8 experts with 7 billion parameters each: “8x7B”... a training method for AI systems in which, instead of a single model learning everything, a combination of sub-models is used", with a gating network " assigning tasks to the experts" and combining their insights.

Test it on

Poe.com and https://app.fireworks.ai/models.

Read the Full Post

The above notes were curated from the full post medium.com/tales-of-tomorrow/a-quiet-revolution-mistral-ai-releases-sensational-new-ai-model-c17c663287f0.

Related reading

More Stuff I Like

More Stuff tagged ai , llm , mixture of experts , mistral

See also: Digital Transformation , Innovation Strategy , Science&Technology , Large language models

Cookies disclaimer

MyHub.ai saves very few cookies onto your device: we need some to monitor site traffic using Google Analytics, while another protects you from a cross-site request forgeries. Nevertheless, you can disable the usage of cookies by changing the settings of your browser. By browsing our website without changing the browser settings, you grant us permission to store that information on your device. More details in our Privacy Policy.