dahara1's Avatar

dahara1

@dahara1.bsky.social

I made machine translation with LLMs. I made PC Chrome translation plugin for bluesky. I made smart feed for bluesky. I mada content import agent. Let's improve these qualities!

90 Followers  |  101 Following  |  416 Posts  |  Joined: 19.08.2023  |  1.7328

Latest posts by dahara1.bsky.social on Bluesky

Preview
OpenAI社に「Sora 2」の運用に関する要望書を提出 | 一般社団法人コンテンツ海外流通促進機構(CODA) CODAは会員社からの要請に基づき、2025年10月27日、OpenAI社に対し、同社が2025年9月30日

@dorialexander.bsky.social

You may not remember, but just in case.

An organization that includes Studio Ghibli has submitted a request to OpenAI's Sora2 not to use it as a learning target without permission.

coda-cj.jp/news/2577/

30.10.2025 10:26 — 👍 1    🔁 0    💬 0    📌 0

The Japanese version of nanochat's middle training will be available soon.

22.10.2025 13:27 — 👍 0    🔁 0    💬 0    📌 0

Unless you're a famous person, your website or blog won't be respected or cited.

They're treated as "data picked up from the internet," and crawlers will just occasionally visit them.

In the age of AI, will we be forced to participate in the attention economy, such as SNS and video streaming?

10.10.2025 05:35 — 👍 0    🔁 0    💬 0    📌 0

I feel like Claude is sometimes better than Gemini at using OSS tools uploaded to GitHub.
Gemini has a lot of hallucinations when it comes to GitHub.

09.10.2025 17:08 — 👍 0    🔁 0    💬 0    📌 0

I chatted with Claude about forgetting some of the settings I'd made in the past, and he automatically searched my chat history.

This is more convenient than searching through chat history.

24.09.2025 03:12 — 👍 0    🔁 0    💬 0    📌 0

That's great, I'd love to make a small model to talk to, but I tend to stray from the main topic.

14.09.2025 12:04 — 👍 1    🔁 0    💬 0    📌 0

Gemini Pro writes the modified code as follows:
# ★ Add from here

# ★ Add up to here

I found it very useful.

However, Be carefuel!
I found that it can be extra work because Gemini comments are sometimes missed.

14.09.2025 11:10 — 👍 0    🔁 0    💬 0    📌 0

LLM Development

Even publicly available foundational models always have some kind of bug.

Even well-known frameworks always have some kind of bug.

Even proven libraries can suddenly stop working.

Datasets always contain some kind of abnormal data.

Conventional wisdom changes at a rapid pace

13.09.2025 05:28 — 👍 0    🔁 0    💬 0    📌 0

What you quickly discover when you challenge an AI agent to search the web.

Google search is truly amazing.

13.09.2025 03:28 — 👍 0    🔁 0    💬 0    📌 0

More and more people are using pre-training or continuous pre-training for SLM, but it's really difficult.

Running it on a consumer-grade GPU isn't hard, but it takes over 1,000 hours. Even then, the data volume is insufficient, resulting in a dumb model.

07.09.2025 09:56 — 👍 0    🔁 0    💬 0    📌 0

EmbeddingGemma was faster than BERT when run on a CPU! BERT was faster on a GPU.

This really seems like it would be good for edge computing.

05.09.2025 08:29 — 👍 0    🔁 0    💬 0    📌 0
Post image

The AI ​​revolution has begun.

04.09.2025 16:47 — 👍 1    🔁 0    💬 0    📌 0

(1)Easy plan
Instruct model + finetune

(2)Solid plan
Base model + Continuous pre training + post training + finetune

(3)Advanced plan
Base model + Continuous pre training + (Instruct model - Base model) + post training + finetune

01.09.2025 12:31 — 👍 0    🔁 0    💬 0    📌 0

If the model places too much emphasis on safety, it will likely start to ignore user instructions.

But who will check the output of an AI agent that doesn't follow instructions?

Will humans be responsible for checking it?

01.09.2025 04:20 — 👍 0    🔁 0    💬 0    📌 0

gpt-oss is fast, but it seems to have poor ability to follow detailed instructions.

Is it better to use it assuming the use of tools?

31.08.2025 02:48 — 👍 0    🔁 0    💬 0    📌 0
Post image

Surprisingly, Gemini didn't get the joke just by looking at it, but Claude did.

The title is "Italians are furious."

26.08.2025 15:00 — 👍 1    🔁 0    💬 0    📌 0
Post image

There are rumors that llama 4.1 and 4.2 are SLM.

15.08.2025 08:58 — 👍 0    🔁 0    💬 0    📌 0

An increasing number of services and products are setting up "AI support chatbots" without publishing product manuals or usage instructions on their websites.

Without reliable documentation, the AI's responses will be hallucinatory and completely useless.

12.08.2025 04:49 — 👍 1    🔁 0    💬 0    📌 0

As AI has made writing easier, contests and other events have begun requiring the submission of explanatory videos.

Overall, it seems like there's more work for humans to do than ever before.

07.08.2025 02:53 — 👍 0    🔁 0    💬 0    📌 0

Opus can no longer do the tasks that it was able to do a few weeks ago. I'm very sad.

02.08.2025 10:40 — 👍 0    🔁 0    💬 0    📌 0

AI winter is coming.

I introduced VoiceCore, LLM-based Japanese TTS I created, to the Japanese-learning community, but reaction was negative.

People are getting tired of the innovative AI-powered learning materials introduced by influencers on TikTok, and anything with the word "AI" is boring.

31.07.2025 06:53 — 👍 0    🔁 0    💬 0    📌 0
Post image 27.07.2025 13:47 — 👍 0    🔁 0    💬 0    📌 0
VoiceCore - AI音声生成システム

We have finally completed a TTS model that can generate emotional Japanese speech from text.

Those who can speak Japanese might be interested.

webbigdata.jp/voice-ai-age...

26.07.2025 07:06 — 👍 0    🔁 0    💬 0    📌 0

Opus omits just two lines of main, saying "the rest of the code is the same"

I spent two hours debugging with Gemini to find out why the app suddenly stopped working

Rage with nowhere to go

25.07.2025 14:02 — 👍 0    🔁 0    💬 0    📌 0

It's hard to find a single prompt that will always give you the perfect answer.

You might want to consider splitting the answer and the verification into two prompts.

25.07.2025 05:17 — 👍 0    🔁 0    💬 0    📌 0
ME: Ask AI to create a fully automated script
AI: AI demands manual pre-work

ME: Ask AI to create a fully automated script AI: AI demands manual pre-work

23.07.2025 06:09 — 👍 0    🔁 0    💬 0    📌 0
Preview
Subliminal Learning: Language models transmit behavioral traits via hidden signals in data We study subliminal learning, a surprising phenomenon where language models transmit behavioral traits via semantically unrelated data. In our main experiments, a "teacher" model with some trait T (su...

Subliminal Learning

The teacher model is given a system prompt to make it like owls.

Instruct it to output about 10 three-digit numbers, and create 10,000 data that are just numbers.

The student model learns this.

For some reason, the student model begins to like owls.
arxiv.org/abs/2507.14805

23.07.2025 03:44 — 👍 0    🔁 0    💬 0    📌 0
Preview
HuggingFaceTB/SmolLM3-3B-checkpoints · Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

SmolLM3-3B-checkpoints

Hugging Face's powerful 3B model (multi-language, up to 128K context expansion) SmolLM3 training checkpoints and loss logs are released

It's quite a large scale, with 11T tokens training on 384 H100s, so I'm grateful for the reference.
huggingface.co/HuggingFaceT...

22.07.2025 02:19 — 👍 1    🔁 0    💬 0    📌 0

I tried QAT(quantized-aware-training) for the first time, but the model's performance was lower than I expected. Is there any trick to it that's different from regular training?

19.07.2025 02:37 — 👍 0    🔁 0    💬 0    📌 0

Diffusion models (Dream 7B) support for llama.cpp has been merged (PR14644)

It's still slow at the moment, but I was impressed that the diffusion language model works properly with my CPU.

16.07.2025 18:05 — 👍 0    🔁 0    💬 0    📌 0

@dahara1 is following 20 prominent accounts