Published onOctober 4, 2024LLMs for everything, all aboard the hype train?llmt5decoderencoder-decoderIs an LLM always the right choice for text generation? (probably not)
Published onSeptember 13, 2024What George Washington taught me about writing with AIllmragiclwritingassistantLLM has the potential to be a super writing assistant, but why does it often fall short?
Published onJuly 25, 2024Survey of Current Modified Transformer Attention DesignsllmattentionRecent Advancements in Attention Mechanism for Long-Sequence Understanding and Generation
Published onJune 28, 2024AI now stands for Apple IntelligencellmappleloraSome exciting details from Apple announcements at WWDC 2024
Published onJune 6, 2024Picking the right LLM deployment configuration in AWSllmawsAWS offers several options for LLM usage, which will you choose?
Published onMay 16, 2024Is your data more or less important in the age of LLMs?llmdataLLMs promise high levels of performance without the need for task specific training, but that doesn't mean that custom datasets are unnecessary.
Published onMay 3, 2024How to Deploy Bedrock Agents using AWS CDKllmagentsbedrockLast week a bedrock agent was deployed manually in AWS Console. Let's now do it the smart way, with Infrastructure as Code using AWS CDK
Published onApril 26, 2024AI Agents with AWS Bedrock (Claude-3 Haiku LLM)llmagentsbedrockA few weeks ago, I used Google's Gemini LLM for function calling. let's see how it works with AWS Bedrock and Anthropic's Claude
Published onMarch 25, 2024Function Calling with Google's Gemini LLMllmagentsgeminiWhen you need an LLM to access new information, LLM function calling is here to help. Let's explore how to use the Gemini function calling to access real information.
Published onMarch 15, 2024Langchain: An Overview of my Experience with the LLM FrameworkllmraglangchainLangchain seems to be a popular choice for developing LLM applications, but when extending it beyond basic use cases, issues emerge.
Published onMarch 8, 2024Why LLMs mess up the simple thingsllmtokenizationchatgptA foundational problem with LLMs: a takeaway from Andrej Karpathy’s lecture on GPT Tokenization.
Published onFebruary 23, 2024Whose Bias Do You Want?llmgeminigoogleThe release of Gemini 1.5 Pro, a glimpse into the difficulty in fine-tuning LLM guardrails
Published onJanuary 18, 2024Fantastic ML/AI and Where to Find Them: A Guide to the Best ML/AI ResourcesLLMSo much is happening in AI, how do you keep up? Here are some of the best resources to follow to stay up to date on AI research and applications.
Published onJanuary 12, 2024Exploration of Inference Speed of 🤗 Transformers Auto-Regressive ModelengineeringLLMAre there easy ways to improve the inference speed of an auto-regressive model?
Published onDecember 22, 2023Can we trust ChatGPT integrations into consumer sites?promptgptllmEarly adopters of LLM integrations risk unexpected outcomes
Published onDecember 15, 2023Diving deep into Mistral AI and their newest model, Mixtral-8x7BmistralgptllmmixtralLooking at the source code to see what makes Mixtral-8x7B so powerful
Published onDecember 8, 2023The most important AI development of 2023 was maybe not GPT-4llamagptllmIn a year full of AI news and hype, which events are going to have the biggest enduring impact on AI?
Published onDecember 1, 2023Automation Bias in the use of Large Language ModelsawsllmsafetyGenerative AI is showing up everywhere: how do we avoid trusting it more than we should?
Published onNovember 10, 2023Who am I? How to give an LLM access to new datadatallmHow to use Retrieval Augmented Generation (RAG) to connect LlaMA-2 with a large number of internet sources
Published onNovember 3, 2023AI Safety: Snippets of Biden's Executive Order on AIsafetyregulationllmOne application of the recent Executive Order from the White House
Published onOctober 27, 2023Using Google's Bard LLM for Rapid Web DevelopmentLLMweb-developmentbardUsing Google's Bard LLM for Creating a Flask and React Recipe Review Application
Published onOctober 20, 2023Training in Sagemaker with Huggingface, Tips & TricksLLMtransformerssagemakerSome hidden tips and tricks to get the most out of sagemaker training jobs
Published onOctober 13, 2023The New Huggingface Templates for Chat Models FeatureLLMtransformerstokenizationtemplatesUpgrade your code with chat model templates
Published onAugust 28, 2023Train LlaMA-2 LLM on your own emails, Part 2. Model TrainingLLMtransformerstokenizationemailpersonalizationTrain an LLM to answer emails for you
Published onAugust 25, 2023Train a LlaMA-2 LLM for free to be an email autocomplete assistant using your own email data! Part 1. Introduction and Data PreparationLLMtransformersLlaMA-2personalizationPrepare a dataset to train an LLM to answer emails for you
Published onAugust 2, 2023End of Sequence Token ExplainedLLMtransformerstokenizationHow is the pad token handled in training a transformer and what's the impact of setting the pad token to be the same as the eos token?