Hacker Newsnew | past | comments | ask | show | jobs | submit | lgrammel's commentslogin

In case you're using JS/TS, I've been working on an abstraction layer for integrating AI models applications, unifying the API for common operations such as text streaming, object generation, and tool usage: https://github.com/lgrammel/modelfusion


So, you essentially took my project ( https://github.com/rubberduck-ai/rubberduck-vscode ), and instead of contributing to it or forking it, you recommitted all the code under your own name, with a few small tweaks.

To be fair, I saw that you give credit, but it's kind of disappointing still given that prob 98% of the project is the work of the Rubberduck contributors. I know Rubberduck is not that active any more, but it's still disappointing.


Shout out to OP for ModelFusion. Probably one of the better ways to use LLMs instead of just integrating the OAI API and reinventing the rest of the wheel.

Handles lots of corner cases and lets you swap in other models easily.

https://github.com/lgrammel/modelfusion


Thanks!


They should include the original copyright notice in their version of the software according to your license.


Credit aside, if this is mostly your code there may be a copyright/license issue.

The original license: https://github.com/rubberduck-ai/rubberduck-vscode/blob/f11a...

The new project: https://github.com/srikanth235/privy/blob/9b4f8ce7e176ab45d5... no mention of the original copyright


Hi @lgrammel, my intention was not to undermine the contributions made by you and others. Added all the original missing contributor's list to the repo. The commit link is http://tinyurl.com/2uzdefak.


Thanks, I appreciate it! Good luck with the project!


I am genuinely curious to know, what is your expectation here? He gave you credit in Github. He called it forever "indebted", which IMO is too much. You put out code under MIT license. MIT license means anyone can take the copy and fork it.

Why are you disappointed? Are you disappointed because a person of Indian origin did it better than you Or are you disappointed that because you wanted all fame but he took your code fork (still valid under MIT license) and did a better marketing than you with wrapper.

If you don't want people to use your use, put behind paywall. It is not difficult to understand.


Considering both repositories are on Github, actually forking it instead of re-commiting it all as 'init commit' would be a good start.

As you said, nothing compels srikanth235 to do this though, but it's a generally more respectable (acceptable?) way to continue someone elses work as a new project.

Also not sure why you had to bring the race of srikanth235 into it.


[flagged]


You need to step back, cool down and get some perspective. No rational person on the planet could interpret his critique as having any racism or even racist undertones.

srikanth235 made the right move by acknowledging the contributor list from the previous project that was lost due to this repo not being a fork of the original.


I second this call to step back and get some perspective, and mainly because no comment by the claimant here was racist, and none of the comment in this chain contained any racist undertones. Rather it seems that our green text friend here has jumped to a conclusion and inferred something that wasn't present.


Why are you bringing nationality politics into this when it's clearly not relevant?


My expectation would have been a fork, or a repo thats started with the full commit history.


Hey!

I've created "StoryTeller", a multi-modal app that quickly generates audio stories for pre-school kids.

StoryTeller is built with the following libraries:

- Fastify - Next.js - shadcn/ui - ModelFusion - Zod

The following AI APIs are used:

- OpenAI (story generation, embeddings) - Eleven Labs (tts) - Lmnt (tts) - thanks for sponsoring the project with credits! - Stability (images)


https://github.com/lgrammel/storyteller

StoryTeller is an exploratory web application that creates short audio stories for pre-school kids.

It used speech-to-text, llms, text-to-speech, embeddings and image generation.

StoryTeller is built with the following libraries:

ModelFusion Fastify Next.js shadcn/ui Zod

The following AI APIs are used:

OpenAI Eleven Labs Lmnt Stability


I think it is more nuanced. This article for example contains examples that suggest otherwise if you want to increase quality (which is a major concern when putting things in production):

https://www.anyscale.com/blog/fine-tuning-llama-2-a-comprehe...


I’m working on ModelFusion, a TypeScript library for working with AI models (llm, image, etc.)

https://github.com/lgrammel/modelfusion

It is only getting limited traction so I’m wondering if I’m missing something fundamental with the approach that I’m taking.


I had it starred! It looks like a nice set of tools for building a multi-modal AI app - I'll give it a try when I flesh out a discord AI gaming app I was working on. Is the multi-modal aspect (image, audio, language) the main focus? Maybe putting that a bit higher in the readme would help it stick - the intro section was a bit too dense and I ended up skimming it.

As for traction, I wonder if there just isn't much interest in AI with JS/TS right now, for whatever reason?


Yes, good point, JS/TS is definitely behind Python. That might explain some of it.

I expect most models to become multi modal in the future and am building towards. A lot of the core logic of agents will nevertheless be text based imo, so that’s a central piece, but I already added text to image and speech to text, and plan to add text to speech next.


This looks awesome! Keep sharing. I think many newbies are looking exactly for something like that including me to get startet with AI :)


Thanks :) I’ll post here when I have larger updates.


If you work with JS or TS, check out this alternative that I've been working on:

https://github.com/lgrammel/modelfusion

It lets you stay in full control over the prompts and control flow while make a lot of things easier and more convenient.


You can also spin up the llama.cpp server ( https://github.com/ggerganov/llama.cpp/tree/master/examples/... ) and e.g. use ModelFusion to generate text, stream text, tokenize, etc. : https://modelfusion.dev/integration/model-provider/llamacpp


llama.cpp has a server with a REST API that you can use: https://github.com/ggerganov/llama.cpp/tree/master/examples/...


ModelFusion is a library for building AI apps, chatbots, and agents. It provides abstractions for AI models, vector indices, and tools.

- Type inference and validation: ModelFusion uses TypeScript and Zod to infer types wherever possible and to validate model responses.

- Flexibility and control: AI application development can be complex and unique to each project. With ModelFusion, you have complete control over the prompts and model settings, and you can access the raw responses from the models quickly to build what you need.

- No chains and predefined prompts: Use the concepts provided by JavaScript (variables, functions, etc.) and explicit prompts to build applications you can easily understand and control. Not black magic.

- More than LLMs: ModelFusion supports other models, e.g., text-to-image and voice-to-text, to help you build rich AI applications that go beyond just text.

- Integrated support features: Essential features like logging, retries, throttling, tracing, and error handling are built-in, helping you focus more on building your application.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: