Hasan Basri AKIRMAK・アキラマキ (白い川)(@HBAkirmak) 's Twitter Profileg
Hasan Basri AKIRMAK・アキラマキ (白い川)

@HBAkirmak

Amazonian @ AWS. Business technologist helping transform companies & industries. SaaS Co-founder. CS MSc, Exec-MBA. Tweets r my own. ML papers/blogs/code 日本語勉強中

ID:18078238

calendar_today12-12-2008 15:38:43

4,6K Tweets

668 Followers

358 Following

Jeff Dean (@🏡)(@JeffDean) 's Twitter Profile Photo

We're starting to roll out API support for Gemini 1.5 Pro for developers. We're excited to see what you build with the 1M token context window!

We'll be onboarding people to the API slowly at first, and then we'll ramp it up. In the meantime, developers can try out Gemini 1.5…

account_circle
Jeff Dean (@🏡)(@JeffDean) 's Twitter Profile Photo

Gemini 1.5 Pro - A highly capable multimodal model with a 10M token context length

Today we are releasing the first demonstrations of the capabilities of the Gemini 1.5 series, with the Gemini 1.5 Pro model. One of the key differentiators of this model is its incredibly long…

Gemini 1.5 Pro - A highly capable multimodal model with a 10M token context length Today we are releasing the first demonstrations of the capabilities of the Gemini 1.5 series, with the Gemini 1.5 Pro model. One of the key differentiators of this model is its incredibly long…
account_circle
Christopher Manning(@chrmanning) 's Twitter Profile Photo

LLMs like ChatGPT are an amazingly powerful breakthrough in AI and a transformative general purpose technology, like electricity or the internet. LLMs will reshape work and our lives this decade. They are not just a blurry photocopier or an extruder of meaningless word sequences.

LLMs like ChatGPT are an amazingly powerful breakthrough in AI and a transformative general purpose technology, like electricity or the internet. LLMs will reshape work and our lives this decade. They are not just a blurry photocopier or an extruder of meaningless word sequences.
account_circle
Philipp Schmid(@_philschmid) 's Twitter Profile Photo

Can we make RAG applications more robust with fine-tuning? A paper by Microsoft and UC Berkley put this to the test to see if small open LLMs, like AI at Meta Llama 7B, can match OpenAI GPT-3.5.
They called it “Retrieval Augmented Fine Tuning (RAFT)”, where you train an LLM…

Can we make RAG applications more robust with fine-tuning? A paper by @Microsoft and UC Berkley put this to the test to see if small open LLMs, like @AIatMeta Llama 7B, can match @OpenAI GPT-3.5. They called it “Retrieval Augmented Fine Tuning (RAFT)”, where you train an LLM…
account_circle
Hasan Basri AKIRMAK・アキラマキ (白い川)(@HBAkirmak) 's Twitter Profile Photo

Trust in the law of cumulative returns.

The power of actively practicing a language 10 Minutes per day results in great outcomes.

I’ve been doing this with DuoLingo for Japanese since 2019. Here are the outcomes.

Trust in the law of cumulative returns. The power of actively practicing a language 10 Minutes per day results in great outcomes. I’ve been doing this with DuoLingo for Japanese since 2019. Here are the outcomes. #japanese #Duolingo #日本語
account_circle
Andrew Ng(@AndrewYNg) 's Twitter Profile Photo

I think AI agentic workflows will drive massive AI progress this year — perhaps even more than the next generation of foundation models. This is an important trend, and I urge everyone who works in AI to pay attention to it.

Today, we mostly use LLMs in zero-shot mode, prompting…

I think AI agentic workflows will drive massive AI progress this year — perhaps even more than the next generation of foundation models. This is an important trend, and I urge everyone who works in AI to pay attention to it. Today, we mostly use LLMs in zero-shot mode, prompting…
account_circle
Hasan Basri AKIRMAK・アキラマキ (白い川)(@HBAkirmak) 's Twitter Profile Photo

huggingface.co/blog/phi2-inte…

run inference on a mid-range laptop powered by an Intel Meteor Lake CPU.

Phi-2 on Intel Meteor Lake

account_circle
Hasan Basri AKIRMAK・アキラマキ (白い川)(@HBAkirmak) 's Twitter Profile Photo

“Nosce te ipsum”
(Know thyself)

There is clearly a magical power & big wisdom in some words.

This is a a beautiful compilation of Latin “Sententia” from a Linguistics Professor Çiğdem Dürüşken in Turkish.

“Nosce te ipsum” (Know thyself) There is clearly a magical power & big wisdom in some words. This is a a beautiful compilation of Latin “Sententia” from a Linguistics Professor Çiğdem Dürüşken in Turkish.
account_circle
Hasan Basri AKIRMAK・アキラマキ (白い川)(@HBAkirmak) 's Twitter Profile Photo

“Peter Thiel believes talented people are not specialists.

Rather, they're almost always polymaths because they're naturally curious.

Curious about businesses, politics, science, history, poetry...or anything that impacts us as humans.”

account_circle
Hasan Basri AKIRMAK・アキラマキ (白い川)(@HBAkirmak) 's Twitter Profile Photo

The Era of 1-bit LLMs:
All Large Language Models are in 1.58 Bits

- significantly more cost-effective in terms of latency, memory, throughput, and energy consumption

- defines a new scaling law and recipe for training: 13B BitNet b1.58 is more efficient, than 3B FP16 LLM.

The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits - significantly more cost-effective in terms of latency, memory, throughput, and energy consumption - defines a new scaling law and recipe for training: 13B BitNet b1.58 is more efficient, than 3B FP16 LLM.
account_circle