Ecoute – a live transcription tool that uses GPT-3.DALLE-E2 – genAI system developed by OpenAI that can create realistic images from a text description.Codex – coding genAI system developed by OpenAI (currently in limited beta mode).Chatsonic – alternative to ChatGPT, powered by GPT-3 but additionally integrates with Google search.ChatGPT – genAI chatbot developed by OpenAI (the current free public version (as at 20/03/23) is still running GPT-3.5 paid ChatGPT Plus accounts have access to the new GPT-4).Canva Docs Magic Write – AI-powered content generation tool (25 free queries in Canva Docs, additional queries with paid version Canva Pro).Bing – Microsoft search engine now with AI-powered functionality requires Microsoft sign-in and using Edge web browser software.Bard – Google's Internet-connected genAI tool, powered by PaLM2 Today, Stability AI announced the launch of Stable Diffusion XL 1.0, a text-to-image model that the company describes as its most advanced release to date.ATLAS.ti – Qualitative data analysis tool now includes a beta version of AI coding powered by the OpenAI GPT model (version not specified).This list includes stand-alone tools plus existing tools that now have genAI incorporated into them. If you are planning to have students use these tools as part of your learning, teaching, and assessment activities, engage your students first in conversations about the tool’s Terms of Use and Privacy Policy and have them provide consent. Reading the tool’s Terms of Use and Privacy Policy is critical for responsible and ethical use. Each tool collects data, ranging from your personal details when signing up for an account, to collecting the data you input into the tool. Please be aware of the data privacy implications of these tools. Please note that inclusion in this list does not indicate endorsement by the University of Wollongong. See for further details.Disclaimer: This list of resources on this page is intended to provide a general sense of the fast-evolving genAI space. The code and the model weights of Whisper are released under the MIT License. The multitask training format uses a set of special tokens that serve as task specifiers or classification targets. All of these tasks are jointly represented as a sequence of tokens to be predicted by the decoder, allowing for a single model to replace many different stages of a traditional speech processing pipeline. Model SizeĪ Transformer sequence-to-sequence model is trained on various speech processing tasks, including multilingual speech recognition, speech translation, spoken language identification, and voice activity detection. Links to both versions are below, check out more details on the Versions page. We still host all other model sizes in a previous version. We’ve created a version of Whisper which only runs the most recent Whisper model, large-v2. It is trained on a large dataset of diverse audio and is also a multi-task model that can perform multilingual speech transcription as well as speech translation and language identification. en models tend to perform better, especially for the tiny.en and base.en models, while the differences would become less significant for the small.en and medium.en models. The authors mention on their GitHub page that for English-only applications, the. Whisper is a general-purpose speech transcription model. Overview of Whisper’s different models (Whisper’s GitHub page).
0 Comments
Leave a Reply. |
AuthorWrite something about yourself. No need to be fancy, just an overview. ArchivesCategories |