#LLM Plugins
Explore tagged Tumblr posts
Text
The Future of Information Gathering with Large Language Models (LLMs)
The Future of Information Gathering with Large Language Models (LLMs)
The Future of Information Gathering: Large Language Models and Their Role in Data Access What’s On My Mind Today? I’ve been doing a lot of thinking about the future, especially about how we find and use information on the internet. I had some ideas about what might change and what it could mean for all of us. To help me flesh out these ideas, I worked with a super-smart 🙂 computer program…
#AI and Digital Transformation#Artificial Intelligence in Data#Data Access with AI#Digital Advertising and AI#Future of Information Gathering#Future of Web Browsing#Internet Privacy and AI#large language models#LLM Plugins#User Control Over Data
0 notes
Text
Master Willem was right, evolution without courage will be the end of our race.
#bloodborne quote but its so appropriate#like big healing church vibes for them tech leaders#who are all shiny eyed diving headfirst into the utopic future#nevermind the fact that the dangers are only rlly obscured by their realism#the dichotomy of this ai will change the world make the world perfect that what we want#and they wanna get it there#but theyre not there#the chatbot is the thing they have to show for#granted when u use plugins and let a bunch of llms do shit together#u see the actual inscrutable magic that they can make happen#and that magic is also so threatening#but were all blinded to it bc we can only rlly acknowledge the watered down simplified realistic view of the reality#which is synival silly and ridiculous bc reality itself is one but so subjective bc we only have each persons interpretation#and its all so socially constructed#so the utopia becomes the dream of these tech leaders#and they wanna make it real#and its almost there#but the nighmare?#the dystopic future#they dont acknowledge#bc they dont dream or try to make it happen#and then theres just the basic reality#but its all tigether#just bc utopia is unlikely but possible#doesnt mean that having the tech for it#will lead to that#bc the system is the same#weve kept inovating non-stop and yeah so much commodification and wuality if life#but not acc changing anytjing
92K notes
·
View notes
Text
Learn about Microsoft Security Copilot
Microsoft Security Copilot (Security Copilot) is a generative AI-powered security solution that helps increase the efficiency and capabilities of defenders to improve security outcomes at machine speed and scale, while remaining compliant to responsible AI principles. Introducing Microsoft Security Copilot: Learn how the Microsoft Security Copilot works. Learn how Security Copilot combines an…
View On WordPress
#AI#assistive copilot#copilot#Defender#Develop whats next#Developer#development#generative AI#Getting Started#incident response#intelligence gathering#intune#investigate#kusto query language#Large language model#llm#Microsoft Entra#natural language#OpenAI#plugin#posture management#prompt#Security#security copilot#security professional#Sentinel#threat#Threat Intelligence#What is New ?
0 notes
Text
Interesting BSD license WP AI plugin..
"SuperEZ AI SEO Wordpress Plugin A Wordpress plugin that utilizes the power of OpenAI GPT-3/GPT-4 API to generate SEO content for your blog or page posts. This Wordpress plugin serves as a personal AI assistant to help you with content ideas and creating content. It also allows you to add Gutenberg blocks to the editor after the assistant generates the content."
g023/SuperEZ-AI-SEO-Wordpress-Plugin: A Wordpress OpenAI API GPT-3/GPT-4 SEO and Content Generator for Pages and Posts (github.com)
#wordpress#wordpress plugins#ai#ai assisted#content creator#content creation#ai generation#wp#blog#ai writing#virtual assistant#llm#app developers#opensource
0 notes
Text
Obsidian And RTX AI PCs For Advanced Large Language Model
How to Utilize Obsidian‘s Generative AI Tools. Two plug-ins created by the community demonstrate how RTX AI PCs can support large language models for the next generation of app developers.
Obsidian Meaning
Obsidian is a note-taking and personal knowledge base program that works with Markdown files. Users may create internal linkages for notes using it, and they can see the relationships as a graph. It is intended to assist users in flexible, non-linearly structuring and organizing their ideas and information. Commercial licenses are available for purchase, however personal usage of the program is free.
Obsidian Features
Electron is the foundation of Obsidian. It is a cross-platform program that works on mobile operating systems like iOS and Android in addition to Windows, Linux, and macOS. The program does not have a web-based version. By installing plugins and themes, users may expand the functionality of Obsidian across all platforms by integrating it with other tools or adding new capabilities.
Obsidian distinguishes between community plugins, which are submitted by users and made available as open-source software via GitHub, and core plugins, which are made available and maintained by the Obsidian team. A calendar widget and a task board in the Kanban style are two examples of community plugins. The software comes with more than 200 community-made themes.
Every new note in Obsidian creates a new text document, and all of the documents are searchable inside the app. Obsidian works with a folder of text documents. Obsidian generates an interactive graph that illustrates the connections between notes and permits internal connectivity between notes. While Markdown is used to accomplish text formatting in Obsidian, Obsidian offers quick previewing of produced content.
Generative AI Tools In Obsidian
A group of AI aficionados is exploring with methods to incorporate the potent technology into standard productivity practices as generative AI develops and speeds up industry.
Community plug-in-supporting applications empower users to investigate the ways in which large language models (LLMs) might improve a range of activities. Users using RTX AI PCs may easily incorporate local LLMs by employing local inference servers that are powered by the NVIDIA RTX-accelerated llama.cpp software library.
It previously examined how consumers might maximize their online surfing experience by using Leo AI in the Brave web browser. Today, it examine Obsidian, a well-known writing and note-taking tool that uses the Markdown markup language and is helpful for managing intricate and connected records for many projects. Several of the community-developed plug-ins that add functionality to the app allow users to connect Obsidian to a local inferencing server, such as LM Studio or Ollama.
To connect Obsidian to LM Studio, just select the “Developer” button on the left panel, load any downloaded model, enable the CORS toggle, and click “Start.” This will enable LM Studio’s local server capabilities. Because the plug-ins will need this information to connect, make a note of the chat completion URL from the “Developer” log console (“http://localhost:1234/v1/chat/completions” by default).
Next, visit the “Settings” tab after launching Obsidian. After selecting “Community plug-ins,” choose “Browse.” Although there are a number of LLM-related community plug-ins, Text Generator and Smart Connections are two well-liked choices.
For creating notes and summaries on a study subject, for example, Text Generator is useful in an Obsidian vault.
Asking queries about the contents of an Obsidian vault, such the solution to a trivia question that was stored years ago, is made easier using Smart Connections.
Open the Text Generator settings, choose “Custom” under “Provider profile,” and then enter the whole URL in the “Endpoint” section. After turning on the plug-in, adjust the settings for Smart Connections. For the model platform, choose “Custom Local (OpenAI Format)” from the options panel on the right side of the screen. Next, as they appear in LM Studio, type the model name (for example, “gemma-2-27b-instruct”) and the URL into the corresponding fields.
The plug-ins will work when the fields are completed. If users are interested in what’s going on on the local server side, the LM Studio user interface will also display recorded activities.
Transforming Workflows With Obsidian AI Plug-Ins
Consider a scenario where a user want to organize a trip to the made-up city of Lunar City and come up with suggestions for things to do there. “What to Do in Lunar City” would be the title of the new note that the user would begin. A few more instructions must be included in the query submitted to the LLM in order to direct the results, since Lunar City is not an actual location. The model will create a list of things to do while traveling if you click the Text Generator plug-in button.
Obsidian will ask LM Studio to provide a response using the Text Generator plug-in, and LM Studio will then execute the Gemma 2 27B model. The model can rapidly provide a list of tasks if the user’s machine has RTX GPU acceleration.
Or let’s say that years later, the user’s buddy is visiting Lunar City and is looking for a place to dine. Although the user may not be able to recall the names of the restaurants they visited, they can review the notes in their vault Obsidian‘s word for a collection of notes to see whether they have any written notes.
A user may ask inquiries about their vault of notes and other material using the Smart Connections plug-in instead of going through all of the notes by hand. In order to help with the process, the plug-in retrieves pertinent information from the user’s notes and responds to the request using the same LM Studio server. The plug-in uses a method known as retrieval-augmented generation to do this.
Although these are entertaining examples, users may see the true advantages and enhancements in daily productivity after experimenting with these features for a while. Two examples of how community developers and AI fans are using AI to enhance their PC experiences are Obsidian plug-ins.
Thousands of open-source models are available for developers to include into their Windows programs using NVIDIA GeForce RTX technology.
Read more on Govindhtech.com
#Obsidian#RTXAIPCs#LLM#LargeLanguageModel#AI#GenerativeAI#NVIDIARTX#LMStudio#RTXGPU#News#Technews#Technology#Technologynews#Technologytrends#govindhtech
3 notes
·
View notes
Text
This Week in Rust 582
Hello and welcome to another issue of This Week in Rust! Rust is a programming language empowering everyone to build reliable and efficient software. This is a weekly summary of its progress and community. Want something mentioned? Tag us at @ThisWeekInRust on X (formerly Twitter) or @ThisWeekinRust on mastodon.social, or send us a pull request. Want to get involved? We love contributions.
This Week in Rust is openly developed on GitHub and archives can be viewed at this-week-in-rust.org. If you find any errors in this week's issue, please submit a PR.
Want TWIR in your inbox? Subscribe here.
Updates from Rust Community
Official
Announcing Rust 1.84.0
This Month in Our Test Infra: December 2024
Foundation
Announcing Rust Global 2025: London
Newsletters
This Month in Rust OSDev: December 2024
Rust Trends Issue #57
Project/Tooling Updates
cargo.nvim - A Neovim plugin for Rust's Cargo commands
Context-Generic Programming Updates: v0.3.0 Release and New Chapters
The RTen machine learning runtime - a 2024 retrospective
Observations/Thoughts
The gen auto-trait problem
Async Rust is about concurrency, not (just) performance
The Emotional Appeal of Rust
[audio] Brave with Anton Lazarev
[audio] Lychee with Matthias Endler
Rust Walkthroughs
Creating an embedded device driver in Rust
Const Evaluation in Rust For Hex Strings Validation
Concurrent and parallel future execution in Rust
[video] Intro to Embassy: embedded development with async Rust
[video] Comprehending Proc Macros
[video] CppCon - C++/Rust Interop: Using Bridges in Practice
Miscellaneous
December 2024 Rust Jobs Report
Tracing Large Memory Allocations in Rust with BPFtrace
On LLMs and Code Optimization
Nand2Tetris - Project 7 (VM Translator Part 1)
Crate of the Week
This week's crate is vidyut, a Sanskrit toolkit containing functionality about meter, segmentation, inflections, etc.
Thanks to Arun Prasad for the self-suggestion!
Please submit your suggestions and votes for next week!
Calls for Testing
An important step for RFC implementation is for people to experiment with the implementation and give feedback, especially before stabilization. The following RFCs would benefit from user testing before moving forward:
RFCs
No calls for testing were issued this week.
Rust
Tracking issue for RFC 3695: Allow boolean literals as cfg predicates
Testing steps
Rustup
No calls for testing were issued this week.
If you are a feature implementer and would like your RFC to appear on the above list, add the new call-for-testing label to your RFC along with a comment providing testing instructions and/or guidance on which aspect(s) of the feature need testing.
RFCs
Rust
Rustup
If you are a feature implementer and would like your RFC to appear on the above list, add the new call-for-testing label to your RFC along with a comment providing testing instructions and/or guidance on which aspect(s) of the feature need testing.
Call for Participation; projects and speakers
CFP - Projects
Always wanted to contribute to open-source projects but did not know where to start? Every week we highlight some tasks from the Rust community for you to pick and get started!
Some of these tasks may also have mentors available, visit the task page for more information.
rama - see if improvements can/have-to be made to rama's http open telemetry layer support
rama – add rama to TechEmpower's FrameworkBenchmark
rama – add rama server benchmark to sharkbench
If you are a Rust project owner and are looking for contributors, please submit tasks here or through a PR to TWiR or by reaching out on X (formerly Twitter) or Mastodon!
CFP - Events
Are you a new or experienced speaker looking for a place to share something cool? This section highlights events that are being planned and are accepting submissions to join their event as a speaker.
Rust Week (Rust NL) | Closes on 2024-01-19 | Utrecht, NL | Event on 2025-05-13 & 2025-05-14
Rust Summit | Rolling deadline | Belgrade, RS | Event on 2025-06-07
If you are an event organizer hoping to expand the reach of your event, please submit a link to the website through a PR to TWiR or by reaching out on X (formerly Twitter) or Mastodon!
Updates from the Rust Project
469 pull requests were merged in the last week
add new {x86_64,i686}-win7-windows-gnu targets
arm: add unstable soft-float target feature
-Zrandomize-layout harder. Foo<T> != Foo<U>
best_blame_constraint: Blame better constraints when the region graph has cycles from invariance or 'static
mir_transform: implement #[rustc_force_inline]
run_make_support: add #![warn(unreachable_pub)]
account for identity substituted items in symbol mangling
add -Zmin-function-alignment
add default_field_values entry to unstable book
add a list of symbols for stable standard library crates
add an InstSimplify for repetitive array expressions
add inherent versions of MaybeUninit methods for slices
add missing provenance APIs on NonNull
assert that Instance::try_resolve is only used on body-like things
avoid ICE: Account for for<'a> types when checking for non-structural type in constant as pattern
avoid replacing the definition of CURRENT_RUSTC_VERSION
cleanup suggest_binding_for_closure_capture_self diag in borrowck
condvar: implement wait_timeout for targets without threads
convert typeck constraints in location-sensitive polonius
depth limit const eval query
detect mut arg: &Ty meant to be arg: &mut Ty and provide structured suggestion
do not ICE when encountering predicates from other items in method error reporting
eagerly collect mono items for non-generic closures
ensure that we don't try to access fields on a non-struct pattern type
exhaustively handle expressions in patterns
fix ICE with references to infinite structs in consts
fix cycle error only occurring with -Zdump-mir
fix handling of ZST in win64 ABI on windows-msvc targets
implement const Destruct in old solver
lower Guard Patterns to HIR
make (unstable API) UniqueRc invariant for soundness
make MIR cleanup for functions with impossible predicates into a real MIR pass
make lit_to_mir_constant and lit_to_const infallible
normalize each signature input/output in typeck_with_fallback with its own span
remove a bunch of diagnostic stashing that doesn't do anything
remove allocations from case-insensitive comparison to keywords
remove special-casing for argument patterns in MIR typeck (attempt to fix perf regression of #133858)
reserve x18 register for aarch64 wrs vxworks target
rm unnecessary OpaqueTypeDecl wrapper
suggest Replacing Comma with Semicolon in Incorrect Repeat Expressions
support target specific optimized-compiler-builtins
unify conditional-const error reporting with non-const error reporting
use a post-monomorphization typing env when mangling components that come from impls
use llvm.memset.p0i8.* to initialize all same-bytes arrays
used pthread name functions returning result for FreeBSD and DragonFly
warn about broken simd not only on structs but also enums and unions when we didn't opt in to it
implement trait upcasting
mir-opt: GVN some more transmute cases
miri: add FreeBSD maintainer; test all of Solarish
miri: added Android to epoll and eventfd test targets
miri: adjust the way we build miri-script in RA, to fix proc-macros
miri: illumos: added epoll and eventfd
miri: supported fioclex for ioctl on macos
miri: switched FreeBSD to pthread_setname_np
miri: use deref_poiner_as instead of deref_pointer
proc_macro: Use ToTokens trait in quote macro
add #[inline] to copy_from_slice
impl String::into_chars
initial fs module for uefi
hashbrown: added Allocator template argument for rustc_iter
account for optimization levels other than numbers
cargo: schemas: Fix 'metadata' JSON Schema
cargo: schemas: Fix the [lints] JSON Schema
cargo: perf: cargo-package: match certain path prefix with pathspec
cargo: fix: emit warnings as warnings when learning rust target info
cargo: make "C" explicit in extern "C"
cargo: setup cargo environment for cargo rustc --print
cargo: simplify SourceID Ord/Eq
rustdoc-json: include items in stripped modules in Crate::paths
rustdoc: use import stability marker in display
rustdoc: use stable paths as preferred canonical paths
rustfmt: drop nightly-gating of the --style-edition flag registration
clippy: add new lint unneeded_struct_pattern
clippy: auto-fix slow_vector_initialization in some cases
clippy: do not intersect spans coming from different contexts
clippy: do not look for significant drop inside .await expansion
clippy: do not propose to elide lifetimes if this causes an ambiguity
clippy: do not remove identity mapping if mandatory mutability would be lost
clippy: do not trigger redundant_pub_crate in external macros
clippy: don't emit machine applicable map_flatten lint if there are code comments
clippy: don't suggest to use cloned for Cow in unnecessary_to_owned
clippy: fix type suggestion for manual_is_ascii_check
clippy: improve needless_as_bytes to also detect str::bytes()
clippy: new lint: manual_ok_err
clippy: remove unneeded parentheses in unnecessary_map_or lint output
rust-analyzer: add a new and improved syntax tree view
rust-analyzer: add config setting which allows adding additional include paths to the VFS
rust-analyzer: re-implement rust string highlighting via tool attribute
rust-analyzer: fix JSON project PackageRoot buildfile inclusion
rust-analyzer: do not compute prettify_macro_expansion() unless the "Inline macro" assist has actually been invoked
rust-analyzer: do not offer completions within macro strings
rust-analyzer: fix env/option_env macro check disregarding macro_rules definitions
rust-analyzer: fix ref text edit for binding mode hints
rust-analyzer: fix a bug with missing binding in MBE
rust-analyzer: fix actual token lookup in completion's expand()
rust-analyzer: fix another issue with fixup reversing
rust-analyzer: fix diagnostics not clearing between flychecks
rust-analyzer: make edition per-token, not per-file
rust-analyzer: implement #[rust_analyzer::skip] for bodies
rust-analyzer: implement implicit sized bound inlay hints
rust-analyzer: improve hover module path rendering
Rust Compiler Performance Triage
A quiet week with little change to the actual compiler performance. The biggest compiler regression was quickly recognized and reverted.
Triage done by @rylev. Revision range: 0f1e965f..1ab85fbd
Summary:
(instructions:u) mean range count Regressions ❌ (primary) 0.4% [0.1%, 1.8%] 21 Regressions ❌ (secondary) 0.5% [0.0%, 2.0%] 35 Improvements ✅ (primary) -0.8% [-2.7%, -0.3%] 6 Improvements ✅ (secondary) -10.2% [-27.8%, -0.1%] 13 All ❌✅ (primary) 0.2% [-2.7%, 1.8%] 27
4 Regressions, 3 Improvements, 3 Mixed; 3 of them in rollups 44 artifact comparisons made in total
Full report here
Approved RFCs
Changes to Rust follow the Rust RFC (request for comments) process. These are the RFCs that were approved for implementation this week:
No RFCs were approved this week.
Final Comment Period
Every week, the team announces the 'final comment period' for RFCs and key PRs which are reaching a decision. Express your opinions now.
RFCs
Supertrait item shadowing v2
Tracking Issues & PRs
Rust
remove support for the (unstable) #[start] attribute
fully de-stabilize all custom inner attributes
Uplift clippy::double_neg lint as double_negations
Optimize Seek::stream_len impl for File
[rustdoc] Add sans-serif font setting
Tracking Issue for PathBuf::add_extension and Path::with_added_extension
Make the wasm_c_abi future compat warning a hard error
const-eval: detect more pointers as definitely not-null
Consider fields to be inhabited if they are unstable
disallow repr() on invalid items
Cargo
No Cargo Tracking Issues or PRs entered Final Comment Period this week.
Language Team
No Language Team Proposals entered Final Comment Period this week.
Language Reference
distinct 'static' items never overlap
Unsafe Code Guidelines
No Unsafe Code Guideline Tracking Issues or PRs entered Final Comment Period this week.
New and Updated RFCs
Make trait methods callable in const contexts
RFC: Allow packages to specify a set of supported targets
Upcoming Events
Rusty Events between 2025-01-15 - 2025-02-12 🦀
Virtual
2025-01-15 | Virtual (London, UK) | London Rust Project Group
Meet and greet with project allocations
2025-01-15 | Virtual (Tel Aviv-Yafo, IL) | Code Mavens 🦀 - 🐍 - 🐪
An introduction to WASM in Rust with Márk Tolmács (Virtual, English)
2025-01-15 | Virtual (Vancouver, BC, CA) | Vancouver Rust
Leptos
2025-01-16 | Virtual (Berlin, DE) | OpenTechSchool Berlin + Rust Berlin
Rust Hack and Learn | Mirror: Rust Hack n Learn Meetup
2025-01-16 | Virtual (San Diego, CA, US) | San Diego Rust
San Diego Rust January 2025 Tele-Meetup
2025-01-16 | Virtual and In-Person (Redmond, WA, US) | Seattle Rust User Group
January Meetup
2025-01-17 | Virtual (Jersey City, NJ, US) | Jersey City Classy and Curious Coders Club Cooperative
Rust Coding / Game Dev Fridays Open Mob Session!
2025-01-21 | Virtual (Tel Aviv-Yafo, IL) | Rust 🦀 TLV
Exploring Rust Enums with Yoni Peleg (Virtual, Hebrew)
2025-01-21 | Virtual (Washington, DC, US) | Rust DC
Mid-month Rustful
2025-01-22 | Virtual (Rotterdam, NL) | Bevy Game Development
Bevy Meetup #8
2025-01-23 & 2025-01-24 | Virtual | Mainmatter Rust Workshop
Remote Workshop: Testing for Rust projects – going beyond the basics
2025-01-24 | Virtual (Jersey City, NJ, US) | Jersey City Classy and Curious Coders Club Cooperative
Rust Coding / Game Dev Fridays Open Mob Session!
2025-01-26 | Virtual (Tel Aviv-Yafo, IL) | Rust 🦀 TLV
Rust and embedded programming with Leon Vak (online in Hebrew)
2025-01-27 | Virtual (London, UK) | London Rust Project Group
using traits in Rust for flexibility, mocking/ unit testing, and more
2025-01-28 | Virtual (Dallas, TX, US) | Dallas Rust User Meetup
Last Tuesday
2025-01-30 | Virtual (Berlin, DE) | OpenTechSchool Berlin + Rust Berlin
Rust Hack and Learn | Mirror: Rust Hack n Learn Meetup
2025-01-30 | Virtual (Charlottesville, VA, US) | Charlottesville Rust Meetup
Quantum Computers Can’t Rust-Proof This!
2025-01-30 | Virtual (Tel Aviv-Yafo, IL) | Code Mavens 🦀 - 🐍 - 🐪
Are We Embedded Yet? - Implementing tiny HTTP server on a microcontroller
2025-01-31 | Virtual (Delhi, IN) | Hackathon Raptors Association
Blazingly Fast Rust Hackathon
2025-01-31 | Virtual (Jersey City, NJ, US) | Jersey City Classy and Curious Coders Club Cooperative
Rust Coding / Game Dev Fridays Open Mob Session!
2025-02-01 | Virtual (Kampala, UG) | Rust Circle Kampala
Rust Circle Meetup
2025-02-04 | Virtual (Buffalo, NY, US) | Buffalo Rust Meetup
Buffalo Rust User Group
2025-02-05 | Virtual (Indianapolis, IN, US) | Indy Rust
Indy.rs - with Social Distancing
2025-02-07 | Virtual (Jersey City, NJ, US) | Jersey City Classy and Curious Coders Club Cooperative
Rust Coding / Game Dev Fridays Open Mob Session!
2025-02-11 | Virtual (Dallas, TX, US) | Dallas Rust User Meetup
Second Tuesday
2025-02-11 | Virtual (Tel Aviv-Yafo, IL) | Code Mavens 🦀 - 🐍 - 🐪
Meet Elusion: New DataFrame Library powered by Rust 🦀 with Borivoj Grujicic
Europe
2025-01-16 | Amsterdam, NL | Rust Developers Amsterdam Group
Meetup @ Avalor AI
2025-01-16 | Karlsruhe, DE | Rust Hack & Learn Karlsruhe
Karlsruhe Rust Hack and Learn Meetup bei BlueYonder
2025-01-18 | Stockholm, SE | Stockholm Rust
Ferris' Fika Forum #8
2025-01-21 | Edinburgh, GB | Rust and Friends
Rust and Friends (evening pub)
2025-01-21 | Ghent, BE | Systems Programming Ghent
Tech Talks & Dinner: Insights on Systems Programming Side Projects (in Rust) - Leptos (full-stack Rust with webassembly), Karyon (distributed p2p software in Rust), FunDSP (audio synthesis in Rust)
2025-01-21 | Leipzig, SN, DE | Rust - Modern Systems Programming in Leipzig
Self-Organized Peer-to-Peer Networks using Rust
2025-01-22 | London, GB | Rust London User Group
Rust London's New Years Party & Community Swag Drop
2025-01-22 | Oberursel, DE | Rust Rhein Main
Rust 2024 Edition and Beyond
2025-01-23 | Barcelona, ES | Barcelona Free Software
Why Build a New Browser Engine in Rust?
2025-01-23 | Paris, FR | Rust Paris
Rust meetup #74
2025-01-24 | Edinburgh, GB | Rust and Friends
Rust and Friends (daytime coffee)
2025-01-27 | Prague, CZ | Rust Prague
Rust Meetup Prague (January 2025)
2025-01-28 | Aarhus, DK | Rust Aarhus
Hack Night - Advent of Code
2025-01-28 | Manchester, GB | Rust Manchester
Rust Manchester January Code Night
2025-01-30 | Augsburg, DE | Rust Meetup Augsburg
Rust Meetup #11: Hypermedia-driven development in Rust
2025-01-30 | Berlin, DE | Rust Berlin
Rust and Tell - Title
2025-02-01 | Brussels, BE | FOSDEM 2025
FOSDEM Rust Devroom
2025-02-01 | Nürnberg, DE | Rust Nuremberg
Technikmuseum Sinsheim
2025-02-05 | Oxford, GB | Oxford Rust Meetup Group
Oxford Rust and C++ social
2025-02-12 | Reading, GB | Reading Rust Workshop
Reading Rust Meetup
North America
2025-01-16 | Nashville, TN, US | Music City Rust Developers
Rust Game Development Series 1: Community Introductions
2025-01-16 | Redmond, WA, US | Seattle Rust User Group
January Meetup
2025-01-16 | Spokane, WA, US | Spokane Rust
Spokane Rust Monthly Meetup: Traits and Generics
2025-01-17 | México City, MX | Rust MX
Multithreading y Async en Rust 101 - HolaMundo - Parte 3
2025-01-18 | Boston, MA, US | Boston Rust Meetup
Back Bay Rust Lunch, Jan 18
2025-01-21 | New York, NY, US | Rust NYC
Rust NYC Monthly Meetup
2025-01-21 | San Francisco, CA, US | San Francisco Rust Study Group
Rust Hacking in Person
2025-01-22 | Austin, TX, US | Rust ATX
Rust Lunch - Fareground
2025-01-23 | Mountain View, CA, US | Hacker Dojo
RUST MEETUP at HACKER DOJO | Rust Meetup at Hacker Dojo - Mountain View Rust Meetup Page
2025-01-28 | Boulder, CO, US | Boulder Rust Meetup
From Basics to Advanced: Testing
2025-02-06 | Saint Louis, MO, US | STL Rust
Async, the Future of Futures
Oceania:
2025-02-04 | Auckland, NZ | Rust AKL
Rust AKL: How We Learn Rust
If you are running a Rust event please add it to the calendar to get it mentioned here. Please remember to add a link to the event too. Email the Rust Community Team for access.
Jobs
Please see the latest Who's Hiring thread on r/rust
Quote of the Week
This is a wonderful unsoundness and I am incredibly excited about it :3
– lcnr on github
Thanks to Christoph Grenz for the suggestion!
Please submit quotes and vote for next week!
This Week in Rust is edited by: nellshamrell, llogiq, cdmistman, ericseppanen, extrawurst, U007D, joelmarcey, mariannegoldin, bennyvasquez, bdillo
Email list hosting is sponsored by The Rust Foundation
Discuss on r/rust
2 notes
·
View notes
Text
There's nowhere to go on the public internet where your posts won't be scraped for AI, sorry. Not discord not mastodon not nothing. No photoshop plugin or robots.txt will protect you from anything other than the politest of scrapers. Maybe if people wanted to bring back private BBSes we could do something with that but at this point, if you're putting something online, expect it to be used for LLM training datasets and/or targeted advertisements. 🥳
#does it suck? yes. have i selected 'don't AI me' on the tumblr toggle? yes. will it do anything? no#that said....it's really just a question of scale#most things have been inherently copyable since the invention of the printing press
7 notes
·
View notes
Text
One reason I haven't experimented with the LLMs as much is that the price of the VRAM requirements rapidly jumps off the deep end.
For Stable Diffusion, the image generator, you can mostly get by with 12GB. It's enough to generate some decent size images and also use regional prompting and the controlnet plugin. About $300, probably.
The smallest Falcon LLM, 7b parameters, weighs in around 15GB, or a $450 16GB card.
A 24GB card will run like $1,500.
A 48GB A6000 will run like $4,000.
An 80GB A100 is like $15,000.
3 notes
·
View notes
Text
Introduction to the LangChain Framework
LangChain is an open-source framework designed to simplify and enhance the development of applications powered by large language models (LLMs). By combining prompt engineering, chaining processes, and integrations with external systems, LangChain enables developers to build applications with powerful reasoning and contextual capabilities. This tutorial introduces the core components of LangChain, highlights its strengths, and provides practical steps to build your first LangChain-powered application.
What is LangChain?
LangChain is a framework that lets you connect LLMs like OpenAI's GPT models with external tools, data sources, and complex workflows. It focuses on enabling three key capabilities: - Chaining: Create sequences of operations or prompts for more complex interactions. - Memory: Maintain contextual memory for multi-turn conversations or iterative tasks. - Tool Integration: Connect LLMs with APIs, databases, or custom functions. LangChain is modular, meaning you can use specific components as needed or combine them into a cohesive application.
Getting Started
Installation First, install the LangChain package using pip: pip install langchain Additionally, you'll need to install an LLM provider (e.g., OpenAI or Hugging Face) and any tools you plan to integrate: pip install openai
Core Concepts in LangChain
1. Chains Chains are sequences of steps that process inputs and outputs through the LLM or other components. Examples include: - Sequential chains: A linear series of tasks. - Conditional chains: Tasks that branch based on conditions. 2. Memory LangChain offers memory modules for maintaining context across multiple interactions. This is particularly useful for chatbots and conversational agents. 3. Tools and Plugins LangChain supports integrations with APIs, databases, and custom Python functions, enabling LLMs to interact with external systems. 4. Agents Agents dynamically decide which tool or chain to use based on the user’s input. They are ideal for multi-tool workflows or flexible decision-making.
Building Your First LangChain Application
In this section, we’ll build a LangChain app that integrates OpenAI’s GPT API, processes user queries, and retrieves data from an external source. Step 1: Setup and Configuration Before diving in, configure your OpenAI API key: import os from langchain.llms import OpenAI # Set API Key os.environ = "your-openai-api-key" # Initialize LLM llm = OpenAI(model_name="text-davinci-003") Step 2: Simple Chain Create a simple chain that takes user input, processes it through the LLM, and returns a result. from langchain.prompts import PromptTemplate from langchain.chains import LLMChain # Define a prompt template = PromptTemplate( input_variables=, template="Explain {topic} in simple terms." ) # Create a chain simple_chain = LLMChain(llm=llm, prompt=template) # Run the chain response = simple_chain.run("Quantum computing") print(response) Step 3: Adding Memory To make the application context-aware, we add memory. LangChain supports several memory types, such as conversational memory and buffer memory. from langchain.chains import ConversationChain from langchain.memory import ConversationBufferMemory # Add memory to the chain memory = ConversationBufferMemory() conversation = ConversationChain(llm=llm, memory=memory) # Simulate a conversation print(conversation.run("What is LangChain?")) print(conversation.run("Can it remember what we talked about?")) Step 4: Integrating Tools LangChain can integrate with APIs or custom tools. Here’s an example of creating a tool for retrieving Wikipedia summaries. from langchain.tools import Tool # Define a custom tool def wikipedia_summary(query: str): import wikipedia return wikipedia.summary(query, sentences=2) # Register the tool wiki_tool = Tool(name="Wikipedia", func=wikipedia_summary, description="Retrieve summaries from Wikipedia.") # Test the tool print(wiki_tool.run("LangChain")) Step 5: Using Agents Agents allow dynamic decision-making in workflows. Let’s create an agent that decides whether to fetch information or explain a topic. from langchain.agents import initialize_agent, Tool from langchain.agents import AgentType # Define tools tools = # Initialize agent agent = initialize_agent(tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True) # Query the agent response = agent.run("Tell me about LangChain using Wikipedia.") print(response) Advanced Topics 1. Connecting with Databases LangChain can integrate with databases like PostgreSQL or MongoDB to fetch data dynamically during interactions. 2. Extending Functionality Use LangChain to create custom logic, such as summarizing large documents, generating reports, or automating tasks. 3. Deployment LangChain applications can be deployed as web apps using frameworks like Flask or FastAPI. Use Cases - Conversational Agents: Develop context-aware chatbots for customer support or virtual assistance. - Knowledge Retrieval: Combine LLMs with external data sources for research and learning tools. - Process Automation: Automate repetitive tasks by chaining workflows. Conclusion LangChain provides a robust and modular framework for building applications with large language models. Its focus on chaining, memory, and integrations makes it ideal for creating sophisticated, interactive applications. This tutorial covered the basics, but LangChain’s potential is vast. Explore the official LangChain documentation for deeper insights and advanced capabilities. Happy coding! Read the full article
#AIFramework#AI-poweredapplications#automation#context-aware#dataintegration#dynamicapplications#LangChain#largelanguagemodels#LLMs#MachineLearning#ML#NaturalLanguageProcessing#NLP#workflowautomation
0 notes
Text
Diversions #7: Additional progress
Wow, so much has happened since #6. Where do I begin? Should I try to recap my travels to Portland or to Atlanta for WordCamp or FinCon? Should I summarize all of the progress we’ve made on the addition to our home? Should I write about my recent upgrade to an M4 MacBook Pro? Or, do I focus on the updates that have come to the ActivityPub plugin that has me rethinking how I use my personal website (yet again)?
How about all of it?!
I’ve posted a few photos from Portland, though I have many more photos that I’ve taken on film that I just haven’t developed yet. This weekend I went to do exactly that and hit a roadblock. It would help if I had fresh chemicals around!
I’ve also commented on what happened at WordCamp and wrote some key takeaways from FinCon. So I won’t rehash any of that.
Instead, I’ll focus on the reminder that traveling (in general) and going to events has a positive impact on your outlook, your network, and your business. Not all of them can be directly measured or are reflected on the bottom row a spreadsheet – but you can feel them as you move forward. You may face a challenge and end up thinking of a company or person you met at an event that can help. In short; go to events.
The addition is moving along quickly now. This sort of project seems to be one where the beginning parts; planning, drawing, documenting, etc. are the slow parts. Then, once the work begins, it is difficult to keep up with.
Working from home affords me the ability to jog outside once or twice a day to snap a photo or two of the progress. I’m very happy that I’m not the one doing the work this time.
The foundation and floor base
I know snags are inevitable. In fact, the forecast at the end of this week already looks like it will slow down progress some. But that is how these sorts of things go.
The new lighter, faster 14″ M4 Pro MacBook Pro is really a delight to use. I hit a snag or two when upgrading the software that was Intel-only to their Apple Silicon counterparts (especially Adobe) but other than just a few hiccups — the upgrade/migration process is so easy.
My previous MacBook Pro was no slouch but this one is a screamer. Building projects, running local LLMs, or just having literally every single app open doesn’t seem to phase it. Using it as a laptop, something I was never able to do with my previous computer, is also really fun. I can actually get a lot of work done and the battery life is great!
My last bit of equipment upgrading is going to be a Thunderbolt dock and at least one external SSD. I need my port situation to be a bit cleaner and I’m hoping a dock will free me up to just have a single wire running into the laptop. And, as of today, I’m not backing up using Time Machine so I cannot delay getting a new SSD.
The social media landscape is a mess! Part of me loves the messiness. It is an interesting time and lots of people are experimenting with many things. New platforms, new ideas, new ways of sharing. Gone are the days of only one or two social networks thriving.
But another part of me is frustrated. I don’t really know what I want to do with my personal blog as it relates to the social web. I have the ActivityPub plugin installed, and it works well to distribute my posts onto the fediverse as a first class citizen. But it isn’t (and likely never will be) perfect.
If you’re at all interested in your website joining the fediverse, I recommend giving the plugin a try and learning how it works. This isn’t cross-posting. It makes your website “an ActivityPub endpoint” so that the account is your website (and/or its authors). It is very cool and interesting.
The reason the ActivityPub plugin has me rethinking my personal blog use, is that I can begin to see a future where all of my favorites, boosts, quote posts, etc. live here on my website – rather than my Mastodon account. But for that I’d want to restructure my site’s design a bit (and likely my RSS feed). Of course, if I explore this at all I’ll write about it.
Regardless of how this phase of the social web shakes out I know I’ll be publishing here on my blog.
An edition of Diversions wouldn’t be complete without some links:
Christoph Rauscher’s newsletter – I like how Christoph publishes his newsletter. Lovely.
Live at Delia’s Third Happening – Site Nonsite live album recording.
Vault – By Commercial Type.
Rex Brasher Field Notes – Lovely little film about Brasher’s work as well.
Rachel Binx’s website – Go take a look.
Apple Pay Plates – This is me buying Crocs or fanny packs.
0 notes
Text
Calling LLMs from client-side JavaScript, converting PDFs to HTML + weeknotes
See on Scoop.it - Education 2.0 & 3.0
I’ve been having a bunch of fun taking advantage of CORS-enabled LLM APIs to build client-side JavaScript applications that access LLMs directly. I also span up a new Datasette plugin …
0 notes
Text
I’m going to approach this as though when tumblr user tanadrin says that they haven’t seen anti-AI rhetoric that doesn’t trade in moral panic, that they’re telling the truth and more importantly that they would would be interested in seeing some. My hope is that you will read this as a reasonable reply, but I’ll be honest upfront that I can’t pretend that this isn’t also personal for me as someone whose career is threatened by generative AI. Personally, I’m not afraid that any LLM will ever surpass my ability to write, but what does scare me is that it doesn’t actually matter. I’m sure I will be automated out whether my artificial replacement can write better than me or not.
This post is kind of long so if watching is more your thing, check out Zoe Bee’s and Philosophy Tube’s video essays, I thought these were both really good at breaking down the problems as well as describing the actual technology.
Also, for clarity, I’m using “AI” and “genAI” as shorthand, but what I’m specifically referring to is Large Language Models (like ChatGpt) or image generation tools (like MidJourney or Dall-E). The term “AI” is used for a lot of extremely useful things that don’t deserve to be included in this.
Also, to get this out of the way, a lot of people point out that genAI is an environmental problem but honestly even if it were completely eco-friendly I’d have serious issues with it.
A major concern that I have with genAI, as I’ve already touched on, is that it is being sold as a way to replace people in creative industries, and it is being purchased on that promise. Last year SAG and the WGA both went on strike because (among other reasons) studios wanted to replace them with AI and this year the Animation Guild is doing the same. News is full of fake images and stories getting sold as the real thing, and when the news is real it’s plagiarised. A journalist at 404 Media did an experiment where he created a website to post AI-powered news stories only to find that all it did was rip off his colleagues. LLMs can’t think of anything new, they just recycle what a human has already done.
As for image generation, there are all the same problems with plagiarism and putting human artists out of work, as well as the overwhelming amount of revenge porn people are creating, not just violating the privacy of random people, but stealing the labour of sex workers to do it.
At this point you might be thinking that these aren’t examples of the technology, but how people use it. That’s a fair rebuttal, every time there’s a new technology there are going to be reports of how people are using it for sex or crimes so let’s not throw the baby out with the bathwater. Cameras shouldn’t be taken off phones just because people use them to take upskirt shots of unwilling participants, after all, people use phone cameras to document police brutality, and to take upskirt shots of people who have consented to them.
But what are LLMs for? As far as I can tell the best use-case is correcting your grammar, which tools like Grammarly already pretty much have covered, so there is no need for a billion-dollar industry to do the same thing. I am yet to see a killer use case for image generation, and I would be interested to hear one if you have it. I know that digital artists have plugins at their disposal to tidy up or add effects/filters to images they’ve created, but again, that’s something that already exists and has been used for very good reason by artists working in the field, not something that creates images out of nothing.
Now let’s look at the technology itself and ask some important questions. Why haven’t they programmed the racism out of GPT-3? The answer to that is complicated and the answer is complicated and sort of boils down to the fact that programmers often don’t realise that racism needs to be programmed out of any technology. Meredith Broussard touches on this in her interview for the Black TikTok Strike of 2021 episode of the podcast Sixteenth Minute, and in her book More Than A Glitch, but to be fair I haven’t read that.
Here's another question I have: shouldn’t someone have been responsible for making sure that multiple image generators, including Google’s, did not have child pornography in their training data? Yes, I am aware that people engaging in moral panics often lean on protect-the-children arguments, and there are many nuanced discussions to be had about how to prevent children from being abused and protect those who have been, but I do think it’s worth pointing out that these technologies have been rolled out before the question of “will people generate CSAM with it?” was fully ironed out. Especially considering that AI images are overwhelming the capacity for investigators to stop instances of actual child abuse.
Again, you might say that’s a problem with how it’s being used and not what it is, but I really have to stress that it is able to do this. This is being put out for everyday people to use and there just aren’t enough safeguards that people can’t get around them. If something is going to have this kind of widespread adoption, it really should not be capable of this.
I’ll sum up by saying that I know the kind of moral panic arguments you’re talking about, the whole “oh, it’s evil because it’s not human” isn’t super convincing, but a lot of the pro-AI arguments have about as much backing. There are arguments like “it will get cheaper” but Goldman Sachs released a report earlier this year saying that, basically, there is no reason to believe that. If you only read one of the links in this post, I recommend that one. There are also arguments like “it is inevitable, just use it now” (which is genuinely how some AI tools are marketed), but like, is it? It doesn’t have to be. Are you my mum trying to convince me to stop complaining about a family trip I don’t want to go on or are you a company trying to sell me a technology that is spying on me and making it weirdly hard to find the opt-out button?
My hot take is that AI bears all of the hallmarks of an economic bubble but that anti-AI bears all of the hallmarks of a moral panic. I contain multitudes.
9K notes
·
View notes
Text
The Mistral AI New Model Large-Instruct-2411 On Vertex AI
Introducing the Mistral AI New Model Large-Instruct-2411 on Vertex AI from Mistral AI
Mistral AI’s models, Codestral for code generation jobs, Mistral Large 2 for high-complexity tasks, and the lightweight Mistral Nemo for reasoning tasks like creative writing, were made available on Vertex AI in July. Google Cloud is announcing that the Mistral AI new model is now accessible on Vertex AI Model Garden: Mistral-Large-Instruct-2411 is currently accessible to the public.
Large-Instruct-2411 is a sophisticated dense large language model (LLM) with 123B parameters that extends its predecessor with improved long context, function calling, and system prompt. It has powerful reasoning, knowledge, and coding skills. The approach is perfect for use scenarios such as big context applications that need strict adherence for code generation and retrieval-augmented generation (RAG), or sophisticated agentic workflows with exact instruction following and JSON outputs.
The new Mistral AI Large-Instruct-2411 model is available for deployment on Vertex AI via its Model-as-a-Service (MaaS) or self-service offering right now.
With the new Mistral AI models on Vertex AI, what are your options?
Using Mistral’s models to build atop Vertex AI, you can:
Choose the model that best suits your use case: A variety of Mistral AI models are available, including effective models for low-latency requirements and strong models for intricate tasks like agentic processes. Vertex AI simplifies the process of assessing and choosing the best model.
Try things with assurance: Vertex AI offers fully managed Model-as-a-Service for Mistral AI models. Through straightforward API calls and thorough side-by-side evaluations in its user-friendly environment, you may investigate Mistral AI models.
Control models without incurring extra costs: With pay-as-you-go pricing flexibility and fully managed infrastructure built for AI workloads, you can streamline the large-scale deployment of the new Mistral AI models.
Adjust the models to your requirements: With your distinct data and subject expertise, you will be able to refine Mistral AI’s models to produce custom solutions in the upcoming weeks.
Create intelligent agents: Using Vertex AI’s extensive toolkit, which includes LangChain on Vertex AI, create and coordinate agents driven by Mistral AI models. To integrate Mistral AI models into your production-ready AI experiences, use Genkit’s Vertex AI plugin.
Construct with enterprise-level compliance and security: Make use of Google Cloud’s integrated privacy, security, and compliance features. Enterprise controls, like the new organization policy for Vertex AI Model Garden, offer the proper access controls to guarantee that only authorized models are accessible.
Start using Google Cloud’s Mistral AI models
Google Cloud’s dedication to open and adaptable AI ecosystems that assist you in creating solutions that best meet your needs is demonstrated by these additions. Its partnership with Mistral AI demonstrates its open strategy in a cohesive, enterprise-ready setting. Many of the first-party, open-source, and third-party models offered by Vertex AI, including the recently released Mistral AI models, can be provided as a fully managed Model-as-a-service (MaaS) offering, giving you enterprise-grade security on its fully managed infrastructure and the ease of a single bill.
Mistral Large (24.11)
The most recent iteration of the Mistral Large model, known as Mistral Large (24.11), has enhanced reasoning and function calling capabilities.
Mistral Large is a sophisticated Large Language Model (LLM) that possesses cutting-edge knowledge, reasoning, and coding skills.
Intentionally multilingual: English, French, German, Spanish, Italian, Chinese, Japanese, Korean, Portuguese, Dutch, Polish, Arabic, and Hindi are among the dozens of languages that are supported.
Multi-model capability: Mistral Large 24.11 maintains cutting-edge performance on text tasks while excelling at visual comprehension.
Competent in coding: Taught more than 80 coding languages, including Java, Python, C, C++, JavaScript, and Bash. Additionally, more specialized languages like Swift and Fortran were taught.
Agent-focused: Top-notch agentic features, including native function calls and JSON output.
Sophisticated reasoning: Cutting-edge reasoning and mathematical skills.
Context length: 128K is the most that Mistral Large can support.
Use cases
Agents: Made possible by strict adherence to instructions, JSON output mode, and robust safety measures
Text: Creation, comprehension, and modification of synthetic text
RAG: Important data is preserved across lengthy context windows (up to 128K tokens).
Coding includes creating, finishing, reviewing, and commenting on code. All popular coding languages are supported.
Read more on govindhtech.com
#MistralAI#ModelLarge#VertexAI#MistralLarge2#Codestral#retrievalaugmentedgeneration#RAG#VertexAIModelGarden#LargeLanguageModel#LLM#technology#technews#news#govindhtech
0 notes
Text
New Samsung Galaxy Z Series Products Integrate the Doubao Large Model - Journal Today Online https://www.merchant-business.com/new-samsung-galaxy-z-series-products-integrate-the-doubao-large-model/?feed_id=135754&_unique_id=669955dba45ff On July 17th, Samsung Electronics l... BLOGGER - #GLOBAL On July 17th, Samsung Electronics launched the new generation Galaxy Z series products for the Chinese market. During the event, Samsung Electronics announced a partnership with Volcano Engine to enhance the intelligent assistant and AI visual access of Galaxy Z Fold6 and Galaxy Z Flip6 smartphones by integrating Doubao large models, improving the smartphones’ intelligent application experience. Previously, Samsung had announced deep cooperation with Google Gemini at overseas product launch events, while in China it chose partners such as Volcano Engine as collaborators for large models.In addition to the AI functions such as drawing circles for search, real-time translation, and voice transcription that have been disclosed, at this China region launch event, Samsung specially showcased the capabilities brought by Galaxy AI based on the Doubao large model for two new foldable phones: when users search for travel-related keywords through Bixby voice assistant, Samsung’s Galaxy AI will search and combine high-quality content sources to provide users with the latest online information and deliver it to users in the form of short video content cards.For example, when users travel in a city, Bixby assistant can rely on the massive content sources of Doubao large model content plugins to provide users with information such as attractions, food, hotels, etc., helping users improve their travel plans and check in at every beautiful spot.In addition, by introducing the Doubao large model single-image AI portrait technology, Samsung users only need to upload a single photo to convert it into various new images in different styles such as business, 3D cartoon, cyberpunk. This allows users to change their avatar style anytime and fully meet personalized needs.According to Zhao Wenjie, Vice President of the Volcano Engine ecosystem, the Doubao large-scale model has been serving many businesses within ByteDance and providing services to enterprise customers through the Volcano Engine. Zhao Wenjie said: “The Volcano Engine is constantly exploring generative AI technology in three aspects: better model performance, continuous reduction of model costs, and making business scenarios easier to implement so that more people can use it and inspire more innovative scenarios.Xu Yuanmo, Vice President of User Experience Strategy for Samsung Electronics Greater China Region, stated that in the global market, Samsung is collaborating with internationally renowned companies to build the Samsung smart mobile product ecosystem; in the Chinese market, Samsung is also working with top domestic companies in various fields to refine the most outstanding products.“In the field of AI, we collaborate deeply with the best domestic partners to fully tap into and integrate Samsung’s hardware and system advantages, jointly committed to creating globally leading AI smartphones for consumers,” said Xu Yuanmo.SEE ALSO: Baidu AI Cloud Collaborates with Samsung China: Galaxy AI Integrates ERNIE LLM http://109.70.148.72/~merchant29/6network/wp-content/uploads/2024/07/pexels-photo-19318474.jpeg #GLOBAL - BLOGGER On July 17th, Samsung Electronics launched the new generation Galaxy Z series products for the Chinese market. During the event, Samsung Electronics announced a partnership with Volcano Engine to enhance the intelligent assistant and AI visual access of Galaxy Z Fold6 and Galaxy Z Flip6 smartphones by integrating Doubao large models, improving the smartphones’ intelligent … Read More
0 notes
Text
New Samsung Galaxy Z Series Products Integrate the Doubao Large Model - Journal Today Online - BLOGGER https://www.merchant-business.com/new-samsung-galaxy-z-series-products-integrate-the-doubao-large-model/?feed_id=135753&_unique_id=669955dabe2a5 On July 17th, Samsung Electronics launched the new generation Galaxy Z series products for the Chinese market. During the event, Samsung Electronics announced a partnership with Volcano Engine to enhance the intelligent assistant and AI visual access of Galaxy Z Fold6 and Galaxy Z Flip6 smartphones by integrating Doubao large models, improving the smartphones’ intelligent application experience. Previously, Samsung had announced deep cooperation with Google Gemini at overseas product launch events, while in China it chose partners such as Volcano Engine as collaborators for large models.In addition to the AI functions such as drawing circles for search, real-time translation, and voice transcription that have been disclosed, at this China region launch event, Samsung specially showcased the capabilities brought by Galaxy AI based on the Doubao large model for two new foldable phones: when users search for travel-related keywords through Bixby voice assistant, Samsung’s Galaxy AI will search and combine high-quality content sources to provide users with the latest online information and deliver it to users in the form of short video content cards.For example, when users travel in a city, Bixby assistant can rely on the massive content sources of Doubao large model content plugins to provide users with information such as attractions, food, hotels, etc., helping users improve their travel plans and check in at every beautiful spot.In addition, by introducing the Doubao large model single-image AI portrait technology, Samsung users only need to upload a single photo to convert it into various new images in different styles such as business, 3D cartoon, cyberpunk. This allows users to change their avatar style anytime and fully meet personalized needs.According to Zhao Wenjie, Vice President of the Volcano Engine ecosystem, the Doubao large-scale model has been serving many businesses within ByteDance and providing services to enterprise customers through the Volcano Engine. Zhao Wenjie said: “The Volcano Engine is constantly exploring generative AI technology in three aspects: better model performance, continuous reduction of model costs, and making business scenarios easier to implement so that more people can use it and inspire more innovative scenarios.Xu Yuanmo, Vice President of User Experience Strategy for Samsung Electronics Greater China Region, stated that in the global market, Samsung is collaborating with internationally renowned companies to build the Samsung smart mobile product ecosystem; in the Chinese market, Samsung is also working with top domestic companies in various fields to refine the most outstanding products.“In the field of AI, we collaborate deeply with the best domestic partners to fully tap into and integrate Samsung’s hardware and system advantages, jointly committed to creating globally leading AI smartphones for consumers,” said Xu Yuanmo.SEE ALSO: Baidu AI Cloud Collaborates with Samsung China: Galaxy AI Integrates ERNIE LLM http://109.70.148.72/~merchant29/6network/wp-content/uploads/2024/07/pexels-photo-19318474.jpeg New Samsung Galaxy Z Series Products Integrate the Doubao Large Model - Journal Today Online - #GLOBAL BLOGGER - #GLOBAL
0 notes