/r/MachineLearning

2,932,515 Subscribers

1

[R] Advice on Fine-Tuning Meta's Segment Anything 2 (SAM) Model — Balancing Edge cases with Generalizability

I was working with SAM2 and have been trying to figure out the best way to fine-tune it for my specific use case. A few considerations that I was hoping get some insights on:

  1. Error Correction vs Generalization: If I'm interested in fine-tuning the model to perform better on cases where it went wrong most on, can I retains its performance on the examples it was already doing well on. i.e. still maintaining (or even improving) its prior generalizability? Or should I have enough number of examples it was doing well already on to preserve that performance?
  2. Which Components to Fine-Tune? In terms of the model's architecture, I've seen different advice on whether to fine-tune just the mask decoder, the prompt encoder, or both. In your experience, is fine-tuning just the mask decoder enough to improve performance, or do you need to adjust the prompt encoder as well? Or maybe there's more to it—like the backbone or other parts of the model? Is it computationally too much of a difference? Or are there other downsides/considerations as well?
  3. Real-World Experiences: For those who have fine-tuned SAM before, how has your experience been? Any tips, tricks, or pitfalls I should watch out for? Also, how did you go about preparing your fine-tuning dataset? Any suggestions on balancing the diversity of data vs focusing on edge cases?
0 Comments
2024/11/09
17:31 UTC

4

[D] latent space forecasting of the next frame

Hey people, I'm searching papers or hints for a computer vision task. I have implemented a Vision Transformer for image classification. In the next step I have to implement a predictor on top of the encoder network of the ViT, which predicts from enc(x_t) -> enc(x_t+1). The predictor should predict the embedding of the next frame. my first idea is a MLP head or decoder network. If someone has tackled a similar task, im happy about recommendations. Ty

2 Comments
2024/11/09
15:34 UTC

2

[P] Benchmark or open source supervised datasets with text or image features and real-valued regression target?

For some reason, I can't seem to find any well known benchmark datasets that have text or images as features, and real-valued targets. Any target range is fine ( (0,1), (-infinity, infinity), (0, infinity), etc.) I have found examples with ordinal classification targets (e.g. integer rating from 1-5), but that doesn't serve my purpose.

Does anyone know of any open source supervised ML data that fits this description? Preferably a benchmarked one with a performance leaderboard.

1 Comment
2024/11/09
14:28 UTC

0

[R][D] Pattern Matching != Reasoning: We analyzed 2 distinct paths to make LLMs actually think [Technical Deep Dive]

Lead ML & Cryptographer researcher here. Just wrapped up a study that might piss some people off, but the data doesn't lie: Current LLMs (yes, even GPT-4) are just incredibly sophisticated autocomplete. Here's why that matters.

TL;DR:
* Current LLMs don't actually reason, they pattern match really well
* We identified two promising paths forward: training-time and inference-time enhancements
* PEFT + Chain of Thought prompting together show surprising results
* All research/code will be open-source

https://blog.bagel.net/p/train-fast-but-think-slow

6 Comments
2024/11/09
13:25 UTC

1

[D] Last Week in Medical AI: Top LLM Research Papers/Models (November 2 - November 9, 2024)

Last Week in Medical AI: Top LLM Research Papers/Models (November 2 - November 9, 2024)

Medical AI Paper of the Week:

  • Google presents*: Exploring Large Language Models for Specialist-level Oncology Care*
    • This paper evaluates AMIE, a conversational diagnostic AI system, in breast oncology using 50 synthetic cancer vignettes. Enhanced with web search retrieval and a self-critique pipeline, AMIE outperformed internal medicine trainees and oncology fellows in generating management plans, evaluated using a detailed clinical rubric encompassing case summarization, plan safety, and treatment recommendations.

Medical LLM & Other Models:

  • AutoProteinEngine: Multimodal Protein LLM

    • This paper introduces AutoProteinEngine (AutoPE), an LLM-powered multimodal AutoML framework for protein engineering, enabling biologists without deep learning expertise to interact with DL models using natural language. AutoPE integrates LLMs with AutoML for model selection (sequence and graph modalities), hyperparameter optimization, and automated data retrieval, demonstrating significant performance improvements over traditional methods in two real-world protein engineering tasks. Code is available at:
  • GSCo: Generalist-Specialist AI Collaboration

    • This paper introduces GSCo, a framework for medical image analysis combining Generalist Foundation Models (GFMs) and specialist models. It develops MedDr, the largest open-source medical GFM, and lightweight specialists for downstream tasks.
  • SAM for Lung X-ray Segmentation

    • This paper explores the application of Meta AI's Segment Anything Model (SAM) to chest X-ray analysis for lung segmentation. Using a transfer learning approach with fine-tuning, the study demonstrates improved performance compared to the original SAM, achieving results comparable to state-of-the-art models like U-Net.
  • MEG: Knowledge-Enhanced Medical QA

    • This paper introduces MEG, a parameter-efficient method for augmenting Large Language Models (LLMs) with medical knowledge graphs using a lightweight mapping network. Evaluated on four medical multiple-choice datasets, MEG achieves a 10.2% accuracy improvement over the Mistral-Instruct baseline and 6.7% over specialized models like BioMistral, demonstrating the benefit of knowledge graph integration.

Frameworks and Methodologies:

  • BrainSegFounder: 3D Neuroimage Analysis
  • PASSION: Sub-Saharan Dermatology Dataset
  • Label Critic: Data-First Approach
  • Medprompt Runtime Strategies

Medical LLM Applications:

  • CataractBot: Patient Support System

  • CheX-GPT: X-ray Report Enhancement

  • CardioAI: Cancer Cardiotoxicity Monitor

  • HealthQ: Healthcare Conversation Chain

  • PRObot: Diabetic Retinopathy Assistant

Medical LLMs & Benchmarks:

  • MediQ: Clinical Reasoning Benchmark

  • Touchstone: Segmentation Evaluation

  • Medical LLM Adaptation Progress

  • Fine-Tuning Medical QA Strategies

AI in Healthcare Ethics:

  • Healthcare Robotics with LLMs
  • XAI in Clinical Practice
  • Precision Rehabilitation Framework
  • Multimodal AI Challenges

Full thread in detail : https://x.com/OpenlifesciAI/status/1855207141302473090

0 Comments
2024/11/09
12:20 UTC

5

[P] MiniBoosts: A small collection of boosting algorithms

Hello, everyone.
I wrote a small collection of boosting algorithms in Rust named MiniBoosts.

This is a hobby project, but I would like to improve more.
Any feedback is welcome.

I appreciate your cooperation.

0 Comments
2024/11/09
12:03 UTC

4

[D] Embeddings and docker file - comparison between two libraries - Is there something better than ONNX?

As title said I was wondering if there are some other ways to embedd corpus without using torch. One of the solution I came up with was by using ONNX. I created the images by using the fastembed library from Qdrant and the sentence-transformer library. Using fastembed result in a significant image size reduction.

Question:

Are there other ways (for example modifying the dockerfile or using other libraries) to shrink the docker image even more?

public repo: https://github.com/learning-bos/dockerize-torch-fastembed-sentence-transformer-comparison

2 Comments
2024/11/09
11:36 UTC

0

[D] Has anyone replaced Transformers with fully-connected layers and verified that it performs strictly worse (for training language models)?

Seems an obvious question but such a "data point" would be very helpful to clear our ignorance.

18 Comments
2024/11/09
09:52 UTC

5

[P] Open-Source Text-to-Agent : framework to develop AI agents from YAML files.

Hey guys, wanted to get your feedback on a project I'm developing. I'm building a framework to define AI agents from YAML configuration files. These files encapsulate tasks that need to be done, how they connect etc, while all the rest is abstracted away.

Now the idea is to use LLMs themselves to create those YAML files from a user prompt. Since the config file has all the core logic of the agent and removes all unnecessary details, I think this is the most efficient way to build a text-to-agent framework. Wdyt?

Let me know your thoughts, and have a look at the repo https://github.com/octopus2023-inc/gensphere

Let me know if you want to contribute and make it work.

1 Comment
2024/11/09
04:47 UTC

2

[D] PAKDD 2023 data?

i was looking into the research papers published in PAKDD 2023. From the names of the authors, I can guess that they are Chinese, Korean, or Japanese

I know PAKDD is a double-blind review. But why other people don't submit their work? or if they submit why the number of acceptance is low

I am also Asian, so I am not trying to be racist here. Just wondering why it is like that

1 Comment
2024/11/09
04:37 UTC

9

[D] Simple ML model hosting service?

My job’s looking for a way for ai to help generate plans, I really think a simple multi-variable model should do the trick; just need to find a reliable hosting service that can be built upon however needed. Are there well established ML hosters that are scalable, configurable, all that?

10 Comments
2024/11/09
01:55 UTC

84

[R] Most Time Series Anomaly Detection results are meaningless (two short videos explain why)

Dear Colleagues

Time Series Anomaly Detection (TSAD) is hot right now, with dozens of  papers each year in NeurIPS, SIGKDD, ICML, PVLDB etc.

However, I claim that much of the published results are meaningless, because the uncertainty of the ground truth labels dwarfs any claimed differences between algorithms or amount of claimed improvements.

I have made two 90-second-long videos that make this clear in a visual and intuitive way:

 1)      Why Most Time Series Anomaly Detection Results are Meaningless (Dodgers)

https://www.youtube.com/watch?v=iRN5oVNvZwk&ab_channel=EamonnKeogh

  2)      Why Most Time Series Anomaly Detection Results are Meaningless (AnnGun)

https://www.youtube.com/watch?v=3gH-65RCBDs&ab_channel=EamonnKeogh

As always, corrections and comments welcome.

Eamonn

 EDIT: To be clear, my point is simply to prevent others from wasting time working with datasets with essentially random labels. In addition, we should be cautious of any claims in the literature that are based on such data (and that includes at least dozens of highly cited papers)

For a review of most of the commonly used TSAD datasets, see this file:

https://www.dropbox.com/scl/fi/cwduv5idkwx9ci328nfpy/Problems-with-Time-Series-Anomaly-Detection.pdf?rlkey=d9mnqw4tuayyjsplu0u1t7ugg&dl=0

46 Comments
2024/11/08
23:58 UTC

7

[D] AI-Generated gameworlds based on classic games? (Ex - Spyro)

I was wondering if anyone had any thoughts on how far out something like this might be or how difficult this is. Ever since the advent of the current era of ai/llms, I thought it would be great to somehow be able to feed data from nostalgic games in some form and create some type of system that is able to generate these worlds infinitely - while still being very true to the style and layout/ethos of the worlds/levels from the reference game. I feel like it would just be so wonderful if there was a path to creating some type of 'never-ending' <insert nostalgic game here> instead of being limited to what the devs put out back in the day.

If anyone has any insight or thoughts on this, please let me know :). I work in the AI space, but I integrate the models, and don't do any training or anything on the low level ML side. Also, yes, I'm only think about the gameworlds/levels atm.

3 Comments
2024/11/08
20:49 UTC

28

[D] Training on Petabyte scale datasets

Lets say we have a dataset that is much larger than we have disk storage. For example:

  • Dataset: 1PB
  • Our disk storage: 10TB
  • GPU RAM: 8x80GB (not super relevant to this discussion)

What are the usual approaches to training on something like this? What I can think of intuitively is to do the following in parallel somehow:

- prefetch block n, train on block n-1, delete block n-2 from disk

Lets say we use PyTorch, so we have a PyTorch Dataset that has all the paths to where the data is stored in the cloud. Do we need to write code for the prefetcher/deleter that downloads from the cloud and store on disk and have it run in a separate process, then have a DataLoader for training that just assumes that it can read from disk (because the prefetcher does its job correctly)? Having the DataLoader read from S3 would be bad for GPU utilization, right?

To take a step back, I'm assuming that this is ordinary and often occuring "problem" for every company that trains on large datasets, so I'm skeptical to writing all of this code by myself as I feel like there should be standard out of the box solutions for this, but can't really find anything that matches perfectly.

16 Comments
2024/11/08
18:27 UTC

0

[R] GPU as a service

Hi all, I have a few GPUs left over from mining, and I’m interested in starting a small-scale GPU-as-a-service. My goal is to set up a simple, side income that could help pay off my credit cards, as I already have a primary job.

What steps are needed for getting started with a small-scale GPU-as-a-service business focused on machine learning or AI? Any insights would be greatly appreciated!

Thanks in advance for any advice you can share!

10 Comments
2024/11/08
17:46 UTC

12

[D] What are crazy structures or update rule that might be useful(or not)? Extreme ideas are welcome

Context: I was making what was supposed to be an FP-oriented NN library/framwork on top of JAX (which too was FP-oriented) called z-zephyr on pip. However, I noticed something you could do with it that kinda clunky, if not tedious, with other frameworks.

(please read context)

TLDR; Zephyr turns out to be very good way (at least in my experience) to make structures that are weird. And I recently just added update capabilities so that zephyr doesn't only do structures but updates too.

Disclaimer: You can this with other frameworks, I have tried many of things I will tell below in other frameworks or libraries, and it's just painful for me or i'm just inexperienced with those.

Here are the crazy things that's quick to do in zephyr, that might not be as quick in other frameworks (if it could be done easily in other frameworks more easily, please tell me).

(These are not supposed to be useful, they're supposed to be extreme)

Full Binary Tree as Neural Network

  • edges have an associated weight
  • input is a scalar (could be a batch with JAX vmap, but let's consider 1)
  • output an array of shape (2^n,) where n is the depth of the tree
  • an update rule that takes into account if the weight is a {L}eft or {R}ight branch (i'll keep it simple, but it can easily be anything)

Here is the tree network in zephyr, and how you get the initial params and tags (tag, is the key in params[key]).

    # essentially 4 lines of code
    @flexible
    def tree_net(params, x, n, i=0):
        if i == n-1:
            return [x]
        return (
            tree_net(
                params["branch"]["L"] if i !=n-2 else params, 
                validate(params["weight"]["L"], (1,), uniform) * x, 
                n, 
                i+1) + 
            tree_net(
                params["branch"]["R"] if i !=n-2 else params, 
                validate(params["weight"]["R"], (1,), uniform) * x, 
                n, 
                i+1)
        )

    x = jnp.ones((1,)) # dummy
    N = 4
    params = trace(tree_net, key, x, N)
    tags = get_lineage_tags(params)

assume you had the loss function and gradients and what not, to keep it simple, i'll just update so that the left branch have weights 0, and the rights ones are kept the same.

    def make_left_zero(params, tags): # i left out gradients 
        if tags[-1] == "L":
            return params * 0
        
        return params

    # update the params 
    params = apply_updates(make_left_zero, params, tags)

Other things you could do with zephyr now (I have tried, and the code is easy for me to do and i'm not that great of a coder)

  • multi-layer network and use the depth of the network (via a tag) to calculate updates of parameters
  • tag some weights as "fast" or "slow" and use those tags in updating
  • create an MLP with neurons as Wx+b. Notice that the neuron is a function that is Array -> Scalar. So I could replace each neuron in that MLP, with another MLP whose output is a scalar (array of shape (1,) ). Or replace the neurons in that with any neural network (any function) that is Array -> Scalar.

What architectures/structures with custom updates rules can you think of that are easy to write(pseudo-code/math or description) but possible cumbersome to implement right now?

Please suggest some extreme idea for me to try.

I think zephyr could be the tooling to make those easy to do. I would like to hear your extreme ideas, so I can try to code them zephyr, and if i can't do it without strugling, and if it's something i think is generic enough, I will evolve zephyr to handle it more easily.

PS: The readme doesn't include these yet, since it started as an (normal) NN library.

The link of the repo will be in the comments if you want to check it out.

9 Comments
2024/11/08
16:14 UTC

1

[P] Build MLPs with Drag-and-Drop and Observe Real-Time Changes While Training in Browser

Hi everyone. I built Grada to learn how things work under the hood. It’s an interactive browser tool that lets you observe real-time changes while training a multilayer perceptron, all built from scratch with a custom tensor-based engine.

You can easily construct neural networks with drag and drop and watch how training affects parameters and outputs visually in real time. Grada also includes a handwritten digit recognition feature, letting you interactively test your model by drawing digits and visualizing predictions. It might be a useful educational tool.

You can find the source code and a quick demo gif on GitHub at https://github.com/saliherdemk/Grada, and the live demo is available at https://saliherdemk.github.io/Grada/.

Hope this helps and looking forward to hearing some feedback.

0 Comments
2024/11/08
08:15 UTC

10

[R] Benchmarking Large Language Models with Integer Sequence Generation Tasks

Benchmarking Large Language Models with Integer Sequence Generation Tasks
Daniel O'Malley, Manish Bhattarai, Javier Santos - Los Alamos National Laboratory
This paper presents a novel benchmark where the large language model (LLM) must write code that computes integer sequences from the Online Encyclopedia of Integer Sequences (OEIS), a widely-used resource for mathematical sequences. The benchmark is designed to evaluate both the correctness of the generated code and its computational efficiency. Our benchmark reveals that the o1 series of models outperform other frontier models from OpenAI, Anthropic, Meta, and Google in accuracy and cheating rates across both easy and hard integer sequences. In order to ensure models do not exploit memorized sequence values, we introduce an automated cheating detection mechanism that flags the use of lookup tables and validated this automation against human cheating evaluations. This benchmark provides a meaningful challenge for current LLMs, offering insights into their mathematical reasoning and code writing capabilities, which can guide future research directions and model development in mathematical reasoning and code synthesis.
arXiv:2411.04372 [cs.LG]: https://arxiv.org/abs/2411.04372

https://preview.redd.it/4vvh5s21unzd1.jpg?width=588&format=pjpg&auto=webp&s=c8bece31712d5d6378188c88e14b9f56e477d41f

1 Comment
2024/11/08
11:07 UTC

2

[D] prediction variability for target with statistics for features

Hi. I'm trying to use ML/DL model for predicting variability statistics like min, max, avg, var, with several features same as target.

For example,

  • Input:
    • min, max, average, variance for the number of customer arrivals in a day
    • min, max, average, variance for the number of customer departures in a day
  • Output:
    • min, max, average, variance for the number of waiting customers in a day

I find several papers related to interval or range prediction for various area like wind power, stock price or solar energy, but I think those papers are different to my purpose. Almost every papers are predicting specific constant value based on time series data first, and use statistical method to estimate prediction interval.

I'm trying to find a way for prediction variability of target value with variability of features. My best idea is make each model to predict each statistics, like one model for minimum, other model for average, ... But I think there is a better way to do this. Is there any ML/DL model or other technique/methodology for this purpose?

2 Comments
2024/11/08
09:51 UTC

2

[D] Looking for Advice & Resources on ASD Prediction Using Voice Cues

Hey everyone! I’m working on my final-year project for my Bachelor’s, where I’m trying to predict Autism Spectrum Disorder (ASD) using voice cues. I’ve worked on some basic ML projects and CNNs before, but this is my first time dealing with audio data, and I’ll be collecting samples from young kids with ASD, from toddlers up to age 12.

I could really use some help finding resources to get a solid grasp on signal processing and how to train classification models specifically on audio. Also, if anyone knows of any open datasets in this area (I haven’t had much luck there) or has any advice or resources, I’d be super grateful. Thanks a ton in advance!

0 Comments
2024/11/08
06:59 UTC

2

[D] What tools do you recommend to manage ML data sets and evaluations?

Hello, our company recently decided to expand our ML team from a very small 2 person team to a more serious efforts.

When we were small, we really didnt have a way to manage data sets or evaluations. They were just files checked into a github repo.

But increasingly we find, with multiple ML models (some llm and some not), and many iterations of datasets (some experimental and some not). It's really hard to version them in a meaningful way and be able to compare and analyze them.

We are a large company, so cost is not really an issue. And all our infrastructure is hosted in Azure. If anything, they fear lock in. What is the best platform/tools for this kind of usage?

3 Comments
2024/11/08
05:28 UTC

6

[D] Directions on drug-target interaction prediction

Almost all the papers I have read on DTI do something like this.

  1. Generates target embeddings using PLMs like ESM2
  2. Generates drug embeddings using CLMs like ChemBERTa
  3. Uses a late fusion or some kind of cross modal attention mechanism.

How to do things differently? Can we use something like docking scores as cross modal attention bias?

1 Comment
2024/11/08
05:27 UTC

43

[D] Just how bad is tfds code quality?

I'm trying a new cute architecture on a bunch of the default datasets out there, using Jax since I'm doing live brain surgery, that part works well.

What I'm having a hell of a time with is actually loading the data. I was going for tfds since its 1) old 2) used in production 3) has a million datasets already prepared. I've not used TF since the 2.0 days and everything seems broken? I'm getting warnings and errors whenever I try loading and running through any dataset. Even their documentation has the errors [0] in the tutorial notebooks.

I can't just ignore a whole bunch of errors and warnings when I'm trying to benchmark a new architecture. Is tfds just that bad or am I missing something obvious?

[0] https://www.tensorflow.org/datasets/overview

11 Comments
2024/11/08
03:25 UTC

0

[D] If I just want an inference engine for any given ML task that gives relatively SOTA results, is there anything better than Hugging Face?

For general prototyping purposes, I don't want to have to train or deploy a model, I just want it behind a service already and to provide it with necessary inputs in the request.... what do you guys think?

EDIT: I suppose for more classical ML tasks, there's no real concept of "pre-trained" in the first place, so you can't just get inference for free... does that sound roughly true?

7 Comments
2024/11/08
00:59 UTC

48

[R]: How much is a noisy image worth? 👀

https://arxiv.org/abs/2411.02780

Shows that corrupted images can be almost as useful as clean images for training generative models, assuming that a small initial set of clean images is available.

This could be useful for dataset design/curation: some budget needs to be invested in obtaining a few high-quality samples and then for the rest of the dataset corrupted images should work fine.

https://preview.redd.it/8vk1nwfexizd1.jpg?width=2952&format=pjpg&auto=webp&s=c6f753956e531303f7818de2c5aa5b5b94d9c2da

Abstract:

The quality of generative models depends on the quality of the data they are trained on. Creating large-scale, high-quality datasets is often expensive and sometimes impossible, e.g. in certain scientific applications where there is no access to clean data due to physical or instrumentation constraints. Ambient Diffusion and related frameworks train diffusion models with solely corrupted data (which are usually cheaper to acquire) but ambient models significantly underperform models trained on clean data. We study this phenomenon at scale by training more than 80 models on data with different corruption levels across three datasets ranging from 30,000 to ≈1.3M samples. We show that it is impossible, at these sample sizes, to match the performance of models trained on clean data when only training on noisy data. Yet, a combination of a small set of clean data (e.g. ~10% of the total dataset) and a large set of highly noisy data suffices to reach the performance of models trained solely on similar-size datasets of clean data, and in particular to achieve near state-of-the-art performance. We provide theoretical evidence for our findings by developing novel sample complexity bounds for learning from Gaussian Mixtures with heterogeneous variances. Our theoretical model suggests that, for large enough datasets, the effective marginal utility of a noisy sample is exponentially worse than that of a clean sample. Providing a small set of clean samples can significantly reduce the sample size requirements for noisy data, as we also observe in our experiments.

Paper: https://arxiv.org/abs/2411.02780

Code: https://github.com/giannisdaras/ambient-laws

Huggingface models: https://huggingface.co/giannisdaras?search_models=ambient_laws

12 Comments
2024/11/07
18:33 UTC

10

[N] Super fast and SOTA Visual Tokenizers

Tokenizers are key to successful development of image and video generative models or multimodal LLMs. Compared to generative models, they are underrated. This work presents many tokenizers that are causal supporting both images and videos in both continuous (relevant in diffusion) and discrete (relevant in autoregressive/transformers) spaces

https://github.com/NVIDIA/Cosmos-Tokenizer

0 Comments
2024/11/07
17:17 UTC

137

[D] Do you get to exercise your ML skills often at your job?

I was hired original as an ML engineer/scientist a few years ago. And for the most part my day to day reflected that. But with the boom of LLMs my team seems to solely focus on using a lot of this tech "out of the box", including agentic wrappers. My work has been dumbed down to prompt engineering to force a huge general purpose model into our domain specific use case. The results are acceptable for the most part, not going to lie, but there's still a small proportion of the cases where a fine-tuned model would have won. The leadership does not seem to be interested in fine-tuning or coming up with something original. A lot of the wrappers especially are very raw and force you into the usage of specific patterns and models. But because they are considered "out of the box", that's what's pushed on us to use. I feel like we are trying to fit a cube into a round hole.

32 Comments
2024/11/07
15:22 UTC

15

[P] ML and LLM system design: 500 case studies to learn from (Airtable database)

Hey everyone! Wanted to share the link to the database of 500 ML use cases from 100+ companies that detail ML and LLM system design. The list also includes over 80 use cases on LLMs and generative AI. You can filter by industry or ML use case.

If anyone here is designing an ML system, I hope you'll find it useful!

Link to the database: https://www.evidentlyai.com/ml-system-design

Disclaimer: I'm on the team behind Evidently, an open-source ML and LLM observability framework. We put together this database.

1 Comment
2024/11/07
15:15 UTC

Back To Top