Tumgik
#Research Model
trcreference · 2 days
Text
0 notes
simplyforensic · 1 year
Text
Persistence of Touch DNA: Understanding the Stability and Implications
Touch DNA, the microscopic genetic material left behind by human touch on surfaces has emerged as a valuable source of evidence in forensic investigations. Recent advancements in DNA analysis techniques have enabled forensic scientists to extract and analyze DNA from touch DNA samples, opening up new possibilities for solving crimes. However, touch DNA’s stability and persistence have remained a…
Tumblr media
View On WordPress
0 notes
triptychofvoids · 5 months
Text
Tumblr media Tumblr media
so @vaxxman, about that coffee?
736 notes · View notes
raillue · 2 days
Text
Exo Failsafe design because she deserves to go on adventures with us <33
Tumblr media
189 notes · View notes
stuckinapril · 6 months
Text
I love girls who’re comfortable being feminine in male dominated stem fields. Such is the case for chemistry but it will not stop me from going to lab in a pink fit if that’s what I want
159 notes · View notes
Text
The real AI fight
Tumblr media
Tonight (November 27), I'm appearing at the Toronto Metro Reference Library with Facebook whistleblower Frances Haugen.
On November 29, I'm at NYC's Strand Books with my novel The Lost Cause, a solarpunk tale of hope and danger that Rebecca Solnit called "completely delightful."
Tumblr media
Last week's spectacular OpenAI soap-opera hijacked the attention of millions of normal, productive people and nonsensually crammed them full of the fine details of the debate between "Effective Altruism" (doomers) and "Effective Accelerationism" (AKA e/acc), a genuinely absurd debate that was allegedly at the center of the drama.
Very broadly speaking: the Effective Altruists are doomers, who believe that Large Language Models (AKA "spicy autocomplete") will someday become so advanced that it could wake up and annihilate or enslave the human race. To prevent this, we need to employ "AI Safety" – measures that will turn superintelligence into a servant or a partner, nor an adversary.
Contrast this with the Effective Accelerationists, who also believe that LLMs will someday become superintelligences with the potential to annihilate or enslave humanity – but they nevertheless advocate for faster AI development, with fewer "safety" measures, in order to produce an "upward spiral" in the "techno-capital machine."
Once-and-future OpenAI CEO Altman is said to be an accelerationists who was forced out of the company by the Altruists, who were subsequently bested, ousted, and replaced by Larry fucking Summers. This, we're told, is the ideological battle over AI: should cautiously progress our LLMs into superintelligences with safety in mind, or go full speed ahead and trust to market forces to tame and harness the superintelligences to come?
This "AI debate" is pretty stupid, proceeding as it does from the foregone conclusion that adding compute power and data to the next-word-predictor program will eventually create a conscious being, which will then inevitably become a superbeing. This is a proposition akin to the idea that if we keep breeding faster and faster horses, we'll get a locomotive:
https://locusmag.com/2020/07/cory-doctorow-full-employment/
As Molly White writes, this isn't much of a debate. The "two sides" of this debate are as similar as Tweedledee and Tweedledum. Yes, they're arrayed against each other in battle, so furious with each other that they're tearing their hair out. But for people who don't take any of this mystical nonsense about spontaneous consciousness arising from applied statistics seriously, these two sides are nearly indistinguishable, sharing as they do this extremely weird belief. The fact that they've split into warring factions on its particulars is less important than their unified belief in the certain coming of the paperclip-maximizing apocalypse:
https://newsletter.mollywhite.net/p/effective-obfuscation
White points out that there's another, much more distinct side in this AI debate – as different and distant from Dee and Dum as a Beamish Boy and a Jabberwork. This is the side of AI Ethics – the side that worries about "today’s issues of ghost labor, algorithmic bias, and erosion of the rights of artists and others." As White says, shifting the debate to existential risk from a future, hypothetical superintelligence "is incredibly convenient for the powerful individuals and companies who stand to profit from AI."
After all, both sides plan to make money selling AI tools to corporations, whose track record in deploying algorithmic "decision support" systems and other AI-based automation is pretty poor – like the claims-evaluation engine that Cigna uses to deny insurance claims:
https://www.propublica.org/article/cigna-pxdx-medical-health-insurance-rejection-claims
On a graph that plots the various positions on AI, the two groups of weirdos who disagree about how to create the inevitable superintelligence are effectively standing on the same spot, and the people who worry about the actual way that AI harms actual people right now are about a million miles away from that spot.
There's that old programmer joke, "There are 10 kinds of people, those who understand binary and those who don't." But of course, that joke could just as well be, "There are 10 kinds of people, those who understand ternary, those who understand binary, and those who don't understand either":
https://pluralistic.net/2021/12/11/the-ten-types-of-people/
What's more, the joke could be, "there are 10 kinds of people, those who understand hexadecenary, those who understand pentadecenary, those who understand tetradecenary [und so weiter] those who understand ternary, those who understand binary, and those who don't." That is to say, a "polarized" debate often has people who hold positions so far from the ones everyone is talking about that those belligerents' concerns are basically indistinguishable from one another.
The act of identifying these distant positions is a radical opening up of possibilities. Take the indigenous philosopher chief Red Jacket's response to the Christian missionaries who sought permission to proselytize to Red Jacket's people:
https://historymatters.gmu.edu/d/5790/
Red Jacket's whole rebuttal is a superb dunk, but it gets especially interesting where he points to the sectarian differences among Christians as evidence against the missionary's claim to having a single true faith, and in favor of the idea that his own people's traditional faith could be co-equal among Christian doctrines.
The split that White identifies isn't a split about whether AI tools can be useful. Plenty of us AI skeptics are happy to stipulate that there are good uses for AI. For example, I'm 100% in favor of the Human Rights Data Analysis Group using an LLM to classify and extract information from the Innocence Project New Orleans' wrongful conviction case files:
https://hrdag.org/tech-notes/large-language-models-IPNO.html
Automating "extracting officer information from documents – specifically, the officer's name and the role the officer played in the wrongful conviction" was a key step to freeing innocent people from prison, and an LLM allowed HRDAG – a tiny, cash-strapped, excellent nonprofit – to make a giant leap forward in a vital project. I'm a donor to HRDAG and you should donate to them too:
https://hrdag.networkforgood.com/
Good data-analysis is key to addressing many of our thorniest, most pressing problems. As Ben Goldacre recounts in his inaugural Oxford lecture, it is both possible and desirable to build ethical, privacy-preserving systems for analyzing the most sensitive personal data (NHS patient records) that yield scores of solid, ground-breaking medical and scientific insights:
https://www.youtube.com/watch?v=_-eaV8SWdjQ
The difference between this kind of work – HRDAG's exoneration work and Goldacre's medical research – and the approach that OpenAI and its competitors take boils down to how they treat humans. The former treats all humans as worthy of respect and consideration. The latter treats humans as instruments – for profit in the short term, and for creating a hypothetical superintelligence in the (very) long term.
As Terry Pratchett's Granny Weatherwax reminds us, this is the root of all sin: "sin is when you treat people like things":
https://brer-powerofbabel.blogspot.com/2009/02/granny-weatherwax-on-sin-favorite.html
So much of the criticism of AI misses this distinction – instead, this criticism starts by accepting the self-serving marketing claim of the "AI safety" crowd – that their software is on the verge of becoming self-aware, and is thus valuable, a good investment, and a good product to purchase. This is Lee Vinsel's "Criti-Hype": "taking press releases from startups and covering them with hellscapes":
https://sts-news.medium.com/youre-doing-it-wrong-notes-on-criticism-and-technology-hype-18b08b4307e5
Criti-hype and AI were made for each other. Emily M Bender is a tireless cataloger of criti-hypeists, like the newspaper reporters who breathlessly repeat " completely unsubstantiated claims (marketing)…sourced to Altman":
https://dair-community.social/@emilymbender/111464030855880383
Bender, like White, is at pains to point out that the real debate isn't doomers vs accelerationists. That's just "billionaires throwing money at the hope of bringing about the speculative fiction stories they grew up reading – and philosophers and others feeling important by dressing these same silly ideas up in fancy words":
https://dair-community.social/@emilymbender/111464024432217299
All of this is just a distraction from real and important scientific questions about how (and whether) to make automation tools that steer clear of Granny Weatherwax's sin of "treating people like things." Bender – a computational linguist – isn't a reactionary who hates automation for its own sake. On Mystery AI Hype Theater 3000 – the excellent podcast she co-hosts with Alex Hanna – there is a machine-generated transcript:
https://www.buzzsprout.com/2126417
There is a serious, meaty debate to be had about the costs and possibilities of different forms of automation. But the superintelligence true-believers and their criti-hyping critics keep dragging us away from these important questions and into fanciful and pointless discussions of whether and how to appease the godlike computers we will create when we disassemble the solar system and turn it into computronium.
The question of machine intelligence isn't intrinsically unserious. As a materialist, I believe that whatever makes me "me" is the result of the physics and chemistry of processes inside and around my body. My disbelief in the existence of a soul means that I'm prepared to think that it might be possible for something made by humans to replicate something like whatever process makes me "me."
Ironically, the AI doomers and accelerationists claim that they, too, are materialists – and that's why they're so consumed with the idea of machine superintelligence. But it's precisely because I'm a materialist that I understand these hypotheticals about self-aware software are less important and less urgent than the material lives of people today.
It's because I'm a materialist that my primary concerns about AI are things like the climate impact of AI data-centers and the human impact of biased, opaque, incompetent and unfit algorithmic systems – not science fiction-inspired, self-induced panics over the human race being enslaved by our robot overlords.
Tumblr media
If you'd like an essay-formatted version of this post to read or share, here's a link to it on pluralistic.net, my surveillance-free, ad-free, tracker-free blog:
https://pluralistic.net/2023/11/27/10-types-of-people/#taking-up-a-lot-of-space
Tumblr media
Image: Cryteria (modified) https://commons.wikimedia.org/wiki/File:HAL9000.svg
CC BY 3.0 https://creativecommons.org/licenses/by/3.0/deed.en
288 notes · View notes
snivyartjpeg · 26 days
Text
Tumblr media
more messy rain code furry doodles + design notes because i think abt them too much <3
60 notes · View notes
bigfatbreak · 7 months
Note
make a post featuring ur donation links and patreon!!!
I'm ok currently! I've been really fighting to bulk up my savings so I have enough to afford a dinky little replacement, it was just REALLY bad timing, considering how recently the crisis with the fridge was... im half expecting my roof to be carried off by a vulture tbh
284 notes · View notes
jstor · 1 year
Text
Tumblr media Tumblr media Tumblr media Tumblr media
Soooo... get this: the squids you're looking at right now ARE MADE OF GLASS! They were created by the renowned 19th-century glass artists Leopold and Rudolf Blaschka (father and son!), and you can find 251 images of these incredible pieces on the Cornell Collection of Blaschka Invertebrate Models on JSTOR, open and free to everyone.
728 notes · View notes
Text
Okay so here comes my mindless ramblings about mordern au! Wwx:
Mordern au! Wwx could be so many things honestly, like he could actually be famous genius researcher, professor, lawyer, crime investigator or even a famous model, pop star. Like the range is crazy, he is just so good at everything (except cooking lol). I personally think he would be a innovative scientist and renowned professor cuz' Wwx is naturally curious, creative and innovative.
I am thinking about how his life would have been better if he was born in today's world, like Wwx was honestly 'ahead of his time' kinda guy, they hated him because he was better, cooler and hotter. If he was famous in the mordern au and the whole getting cancelled thing were to happen, it would probably happen because someone tried to frame him *cough cough jgy or Su she cough cough* for the shit he didn't do or something.
But yeah Wwx is born to be an exceptional talent, a prodigy, he is THE HOT GUY EVER. I love him. Thank you reading this ramble.
60 notes · View notes
obstinatecondolement · 11 months
Text
There is an aphorism in science that all models are wrong, but some are useful. The general idea is that a simplified representation of something much more complex may not perfectly replicate every element of the real thing, or account for every single factor that would affect it under real world conditions, but a good simplification potentially can approximate something more complex enough to get broadly accurate* insights that are useful.
In my opinion, specific sexualities and genders (all of them, fwiw), and the even the concept of being cis or trans, are best thought of as useful models for certain amorphous clusters of experiences and feelings, rather than as things that have concrete, inflexible definitions that map perfectly onto every single person who uses that model of identity as a shorthand. Dictionary definitions of what gay means/what a woman is/etc., are all assuming spherical cows in a vacuum to make the maths easier, and you look like an idiot if you think that cows really are spherical and are not affected by atmospheric pressure in any way (or indeed that they could survive vacuum conditions) and then go around harassing cows on this basis.
A person's internal sense of self is more important than your belief in a model. Fuck off and let me get back to chewing cud.
196 notes · View notes
heartru · 5 months
Text
my least favourite thing from the watcher discourse has been people saying “WE never asked for higher quality” or “your most popular shows are where you just sit around and talk!!” - babes its likely not what THEY want to do for the rest of eternity lol. they’re allowed to want to grow as creatives and make things they are proud of?
64 notes · View notes
spitblaze · 1 month
Text
I am living proof of the Gamer Girl to Transmasc pipeline
49 notes · View notes
rosewind2007 · 2 months
Text
A month ago I made a little model Murderbot out of some kiddies play modeling dough/clay:
Tumblr media
It looked kinda lonely, so I made it a friend
Tumblr media
The Lego is a Dune Ornithopter (a hopper? ART’s hopper?)
40 notes · View notes
carpsoup · 4 days
Text
making an astronomy/meteoritics iterator oc when i know fuckall or at least just very basic things about those things was maybe a mistake. Looking up stuff for reference/inspo like haha i like your funny words magic man
24 notes · View notes
semioticapocalypse · 4 months
Photo
Tumblr media
Carl Woese with an RNA model at General Electric Research Laboratory in 1961.
I Am Collective Memories   •    Follow me, — says Visual Ratatosk
23 notes · View notes