A very short summary of what is RISC-V. ⌨ 👁️🗨️ https://lnkd.in/eXCVgEKm
Diego Correa Tristain’s Post
More Relevant Posts
-
Full-stack R&D Scientist and engineer, working on X-ray scattering, materials development and lab automation
Want to know more about how to store automated synthesis data in an HDF5 file structure? Here's an example from our lab: https://lnkd.in/eNDVZpY7
How to structure synthesis metadata intelligibly? An example from 1200 syntheses on RoWaN.
https://lookingatnothing.com
To view or add a comment, sign in
-
If benchmarks are your thing, and you're curious about how Extism can push over 1GiB/s across the #WebAssembly host/guest boundary, then this is the post for you! https://lnkd.in/gJsCkgee 👏 Thanks to Chris Dickinson for the deep dive!
Back of the Napkin Wasm Performance: How Does Extism Work?
dylibso.com
To view or add a comment, sign in
-
A geek who can speak, Co-creator of PiML (Python interpretable Machine Learning), SVP Risk & Technology H2O.ai, Retired EVP-Head of Wells Fargo MRM
📢 Check out the new release of PiML, developed by our team of data scientists! This update includes inherently interpretable machine learning and model validation, complete with tests for weakness detection, robustness, reliability, resilience, and fairness. Install or update now to experience it for yourself. #PiML #MachineLearning #ModelValidation #ModelRiskManagement
Head of Machine Learning and Validation Engineering at Wells Fargo, Co-creator of PiML Toolbox, PhD/Former Professor of Statistics
🚀 Sep 28, 2023 PiML v0.5.1 release with new features: - Model-free diagnostic test APIs. - Sliced overfitting test with ACC and AUC metrics. - Other miscellaneous improvements and bug fixes. - Support for a wider range of computing environments, including Mac ARM. Try it out by "pip install PiML" User Guide: https://lnkd.in/gA7YdzHx #PiML #machinelearning #interpretability #modeltesting #modelriskmanagement
GitHub - SelfExplainML/PiML-Toolbox: PiML (Python Interpretable Machine Learning) toolbox for model development & diagnostics
github.com
To view or add a comment, sign in
-
Pretty cool stuff by Lamini for enterprise use cases where 95% LLM accuracy and no hallucination could be achieved. Better than SOTA LLM + Prompt + RAG? - Memory Tuning tunes a massive mixture of memory experts on any open-source LLM. - Each memory expert acts like a LoRA adapter that functionally operates as memory for the model. - Together, the memory experts specialize in a million different ways to ensure faithful and factual accuracy to the data that it was tuned on. https://lnkd.in/gnMqaacR
Introducing Lamini Memory Tuning: 95% LLM Accuracy, 10x Fewer Hallucinations | Lamini - Enterprise LLM Platform
lamini.ai
To view or add a comment, sign in
-
BPE__BytePairEncoding https://lnkd.in/dwa_P8Xx bigram = min(pairs, key = lambda pair: self.bpe_ranks.get(pair, float('inf'))) https://lnkd.in/dfFFJsyA The only literal byte pair left occurs only once, and the encoding might stop here. Alternatively, the process could continue with recursive byte pair encoding, replacing "ZY" with "X": XdXac X=ZY Y=ab Z=aa
To view or add a comment, sign in
-
RDF Dataset Canonicalization becomes a W3C recommendation. It provides a standard RDF Dataset canonicalization algorithm #rdf #graph #linkeddata #webstandards #knowledgegraphs #graph #database https://lnkd.in/e2WYdt43
RDF Dataset Canonicalization
w3.org
To view or add a comment, sign in
-
If you want to know more about certifying local explanations, follow this work at ICML next week!
I will be presenting our work Trust Regions for Explanations https://lnkd.in/eAfYyeBT next week at ICML. It is a novel problem where we find regions around examples of interest where the explanations for them would be still valid. This has benefits such as i) insight into the model behavior with a guarantee, ii) ascertaining stability of the explanations, iii) explanation reuse which can significantly save on queries to the model and iv) serving as a possible meta-metric to compare explanation methods. Code for this is now available through our AIX360 toolkit https://lnkd.in/e57t6xHj Thanks to my co-authors Swagatam Haldar Dennis Wei Karthikeyan Natesan Ramamurthy and also to Vijay Arya for making this possible.
Trust Regions for Explanations via Black-Box Probabilistic Certification
arxiv.org
To view or add a comment, sign in
-
What is blocking LLMs from using long context length inputs? 🚨Introducing KVQuant which allows serving LLaMA-7B with 1M tokens on a single A100! 🔥 Current largest model is Claude-2.1 which is limited to 200K tokens. What is the challenge for increasing this? Two key problems: (i) Memory wall: the need to store Key and Value activations for each token, leads to a major issue for long context lengths: we quickly go out of memory even on 8 GPU systems. You can’t run LLaMA-70B with >32K on an A100 because of this. (ii) Catastrophic forgetting: LLMs do not pay attention to long sequences and either only focus on the beginning or the end. So far, training on long documents, and better positional encoding has enabled up to 200K context length, but further scaling to longer context lengths without accuracy drop is an open problem. KVQuant addresses the first challenge, and enables large context length inference by quantizing the cached Key/Value activations to ultra low precision without accuracy degradation by considering several consistent patterns observed in cached KV values across different LLMs. Why does this matter? The ability to consume such large context sizes could help increase the accuracy of LLMs by allowing one to provide different in context examples or include large files/contents in a general LLM and make it competitive with a fine-tuned model that is trained on such data. For instance, consider coding co-pilot. 10M context length provides enough space for the LLM to consume ~1M lines of code, which could help unlock new insights. Please see https://lnkd.in/gmtrsK3P for a quick TLDR of our method. Paper: https://lnkd.in/g6y7Yw3x Code: https://lnkd.in/gPmnNAET Joint work with: Coleman Hooper, Sehoon Kim, Hiva Mohammadzadeh, Michael Mahoney, Sophia Shao, Kurt Keutzer
To view or add a comment, sign in
-
Thought Leader in AI and Business Intelligence with customers in 4 continents. Global keynote speaker. Topics: AI and BI for business, Leadership, Digital Transformation, Diversity and Inclusion, Women in Tech
Leveraging phi-3 for an Enhanced Semantic Cache in RAG Applications. #RAG applications utilize a semantic cache layer and phi-3, a #SLM from Microsoft, to address the challenge of repeat queries while maintaining contextually accurate and diverse responses https://zurl.co/vHg3
Leveraging phi-3 for an Enhanced Semantic Cache in RAG Applications
techcommunity.microsoft.com
To view or add a comment, sign in
-
Leveraging phi-3 for an Enhanced Semantic Cache in RAG Applications. #RAG applications utilize a semantic cache layer and phi-3, a #SLM from Microsoft, to address the challenge of repeat queries while maintaining contextually accurate and diverse responses https://zurl.co/vHg3
Leveraging phi-3 for an Enhanced Semantic Cache in RAG Applications
techcommunity.microsoft.com
To view or add a comment, sign in
Rust Developer | Blockchain Architect | Web3 Advisor
10mohttps://riscv.org/