Data Science at Home
Episodes

Monday Nov 25, 2024
Humans vs. Bots: Are You Talking to a Machine Right Now? (Ep. 273)
Monday Nov 25, 2024
Monday Nov 25, 2024
In this episode of Data Science at Home, host Francesco Gadaleta dives deep into the evolving world of AI-generated content detection with experts Souradip Chakraborty, Ph.D. grad student at the University of Maryland, and Amrit Singh Bedi, CS faculty at the University of Central Florida.
Together, they explore the growing importance of distinguishing human-written from AI-generated text, discussing real-world examples from social media to news. How reliable are current detection tools like DetectGPT? What are the ethical and technical challenges ahead as AI continues to advance? And is the balance between innovation and regulation tipping in the right direction?
Tune in for insights on the future of AI text detection and the broader implications for media, academia, and policy.
Chapters
00:00 - Intro
00:23 - Guests: Souradip Chakraborty and Amrit Singh Bedi
01:25 - Distinguish Text Generation By AI
04:33 - Research on Safety and Alignment of Generative Model
06:01 - Tools to Detect Generated AI Text
11:28 - Water Marking
18:27 - Challenges in Detecting Large Documents Generated by AI
23:34 - Number of Tokens
26:22 - Adversarial Attack
29:01 - True Positive and False Positive of Detectors
31:01 - Limit of Technologies
41:01 - Future of AI Detection Techniques
46:04 - Closing Thought
Subscribe to our new YouTube channel https://www.youtube.com/@DataScienceatHome

Wednesday Nov 06, 2024
Love, Loss, and Algorithms: The Dangerous Realism of AI (Ep. 270)
Wednesday Nov 06, 2024
Wednesday Nov 06, 2024
Subscribe to our new channel https://www.youtube.com/@DataScienceatHome
In this episode of Data Science at Home, we confront a tragic story highlighting the ethical and emotional complexities of AI technology. A U.S. teenager recently took his own life after developing a deep emotional attachment to an AI chatbot emulating a character from Game of Thrones. This devastating event has sparked urgent discussions on the mental health risks, ethical responsibilities, and potential regulations surrounding AI chatbots, especially as they become increasingly lifelike.
🎙️ Topics Covered:
AI & Emotional Attachment: How hyper-realistic AI chatbots can foster intense emotional bonds with users, especially vulnerable groups like adolescents.
Mental Health Risks: The potential for AI to unintentionally contribute to mental health issues, and the challenges of diagnosing such impacts. Ethical & Legal Accountability: How companies like Character AI are being held accountable and the ethical questions raised by emotionally persuasive AI.
🚨 Analogies Explored:
From VR to CGI and deepfakes, we discuss how hyper-realism in AI parallels other immersive technologies and why its emotional impact can be particularly disorienting and even harmful.
🛠️ Possible Mitigations:
We cover potential solutions like age verification, content monitoring, transparency in AI design, and ethical audits that could mitigate some of the risks involved with hyper-realistic AI interactions. 👀 Key Takeaways: As AI becomes more realistic, it brings both immense potential and serious responsibility. Join us as we dive into the ethical landscape of AI—analyzing how we can ensure this technology enriches human lives without crossing lines that could harm us emotionally and psychologically. Stay curious, stay critical, and make sure to subscribe for more no-nonsense tech talk!
Chapters
00:00 - Intro
02:21 - Emotions In Artificial Intelligence
04:00 - Unregulated Influence and Misleading Interaction
06:32 - Overwhelming Realism In AI
10:54 - Virtual Reality
13:25 - Hyper-Realistic CGI Movies
15:38 - Deep Fake Technology
18:11 - Regulations To Mitigate AI Risks
22:50 - Conclusion
#AI#ArtificialIntelligence#MentalHealth#AIEthics#podcast#AIRegulation#EmotionalAI#HyperRealisticAI#TechTalk#AIChatbots#Deepfakes#VirtualReality#TechEthics#DataScience#AIDiscussion #StayCuriousStayCritical

Saturday Oct 12, 2024
What Big Tech Isn’t Telling You About AI (Ep. 267)
Saturday Oct 12, 2024
Saturday Oct 12, 2024
Are AI giants really building trustworthy systems? A groundbreaking transparency report by Stanford, MIT, and Princeton says no. In this episode, we expose the shocking lack of transparency in AI development and how it impacts bias, safety, and trust in the technology. We’ll break down Gary Marcus’s demands for more openness and what consumers should know about the AI products shaping their lives.
Check our new YouTube channel https://www.youtube.com/@DataScienceatHome and Subscribe!
Cool links
https://mitpress.mit.edu/9780262551069/taming-silicon-valley/
http://garymarcus.com/index.html

Thursday Aug 08, 2024
Data Guardians: How Enterprises Can Master Privacy with MetaRouter (Ep. 261)
Thursday Aug 08, 2024
Thursday Aug 08, 2024
In this insightful episode, we dive deep into the pressing issue of data privacy, where 86% of U.S. consumers express growing concerns and 40% don't trust companies to handle their data ethically. Join us as we chat with the Vice President of Engineering at MetaRouter, a cutting-edge platform enabling enterprises to regain control over their customer data. We explore how MetaRouter empowers businesses to manage data in a 1st-party context, ensuring ethical, compliant handling while navigating the complexities of privacy regulations.
Sponsors
Intrepid AI (https://intrepid.ai) is an AI assisted all-in-one platform for robotics teams. Build robotics applications in minutes, not months
References
https://www.metarouter.io/post/mastering-data-governance-why-governance-at-the-point-of-collection-is-a-must-have
https://hubs.ly/Q02HwJly0
https://www.metarouter.io/post/why-privacy-sandbox-is-a-good-paradigm-shift-for-consumers

Tuesday Jun 11, 2024
Harnessing AI for Cybersecurity: Expert Tips from QFunction (Ep. 258)
Tuesday Jun 11, 2024
Tuesday Jun 11, 2024
In this episode, we sit down with Ryan Smith, Founder of QFunction LLC, to explore how AI and machine learning are revolutionizing cybersecurity. With over 8 years of experience, including work at NASA's Jet Propulsion Laboratory, Ryan shares insights on the future of threat detection and prevention, the challenges businesses face in maintaining effective cybersecurity, and the ethical considerations of AI implementation. Learn about cost-effective strategies for small businesses, the importance of collaboration in combating cyber threats, and how QFunction tailors its AI solutions to meet diverse industry needs.
Sponsors
Arctic Wolf Learn what the new year holds for ransomware as a service, Active Directory, artificial intelligence and more when you download the 2024 Arctic Wolf Labs Predictions Report today at arcticwolf.com/datascience
Intrepid AI (https://intrepid.ai) is an AI assisted all-in-one platform for robotics teams. Build robotics applications in minutes, not months.
QFunction does cybersecurity differently. By relying on scientific breakthroughs in AI and machine learning, QFunction works within your existing security stack to detect anomalies and threats within your data
References
QFunction Blog
Threat Hunting Firewall Logs
Threat Hunting Linux Logs
QFunction GitHub

Monday Sep 18, 2023
Attacking LLMs for fun and profit (Ep. 239)
Monday Sep 18, 2023
Monday Sep 18, 2023
As a continuation of Episode 238, I explain some effective and fun attacks to conduct against LLMs. Such attacks are even more effective on models served locally, that are hardly controlled by human feedback.
Have great fun and learn them responsibly.
References
https://www.jailbreakchat.com/
https://www.reddit.com/r/ChatGPT/comments/10tevu1/new_jailbreak_proudly_unveiling_the_tried_and/
https://arxiv.org/abs/2305.13860

Tuesday Oct 25, 2022
Private machine learning done right (Ep. 207)
Tuesday Oct 25, 2022
Tuesday Oct 25, 2022
There are many solutions to private machine learning. I am pretty confident when I say that the one we are speaking in this episode is probably one of the most feasible and reliable.I am with Daniel Huynh, CEO of Mithril Security, a graduate from Ecole Polytechnique with a specialisation in AI and data science. He worked at Microsoft on Privacy Enhancing Technologies under the office of the CTO of Microsoft France. He has written articles on Homomorphic Encryptions with the CKKS explained series (https://blog.openmined.org/ckks-explained-part-1-simple-encoding-and-decoding/). He is now focusing on Confidential Computing at Mithril Security and has written extensive articles on the topic: https://blog.mithrilsecurity.io/.
In this show we speak about confidential computing, SGX and private machine learning
References
Mithril Security: https://www.mithrilsecurity.io/
BindAI GitHub: https://github.com/mithril-security/blindai
Use cases for BlindAI:Deploy Transformers models with confidentiality: https://blog.mithrilsecurity.io/transformers-with-confidentiality/
Confidential medical image analysis with COVID-Net and BlindAI: https://blog.mithrilsecurity.io/confidential-covidnet-with-blindai/
Build a privacy-by-design voice assistant with BlindAI: https://blog.mithrilsecurity.io/privacy-voice-ai-with-blindai/
Confidential Computing Explained: https://blog.mithrilsecurity.io/confidential-computing-explained-part-1-introduction/
Confidential Computing Consortium: https://confidentialcomputing.io/
Confidential Computing White Papers: https://confidentialcomputing.io/white-papers-reports/
List of Intel processors with Intel SGX:https://www.intel.com/content/www/us/en/support/articles/000028173/processors.html
https://github.com/ayeks/SGX-hardware
Azure Confidential Computing VMs with SGX:Azure Docs: https://docs.microsoft.com/en-us/azure/confidential-computing/confidential-computing-enclaves
How to deploy BlindAI on Azure: https://docs.mithrilsecurity.io/getting-started/cloud-deployment/azure-dcsv3
Confidential Computing 101: https://www.youtube.com/watch?v=77U12Ss38Zc
Rust: https://www.rust-lang.org/
ONNX: https://github.com/onnx/onnx
Tract, a Rust inference engine for ONNX models: https://github.com/sonos/tract

Tuesday Dec 14, 2021
Capturing Data at the Edge (Ep. 180)
Tuesday Dec 14, 2021
Tuesday Dec 14, 2021
In this episode I speak with Manavalan Krishnan from Tsecond about capturing massive amounts of data at the edge with security and reliability in mind.
This episode is brought to you by Tsecond
The growth of data being created at static and moving edges across industries such as air travel, ocean and space exploration, shipping and freight, oil and gas, media, and more proposes numerous challenges in capturing, processing, and analyzing large amounts of data.
and by Amethix Technologies
Amethix use advanced Artificial Intelligence and Machine Learning to build data platforms and predictive engines in domain like finance, healthcare, pharmaceuticals, logistics, energy. Amethix provide solutions to collect and secure data with higher transparency and disintermediation, and build the statistical models that will support your business.
References
https://tsecond.us/company/manavalan-krishnan/

Tuesday Jun 15, 2021
Time to take your data back with Tapmydata (Ep. 156)
Tuesday Jun 15, 2021
Tuesday Jun 15, 2021
In this episode I am with Gilbert Hill, head of strategy at https://tapmydata.com/
We speak about personal data, blockchain and the ability to control it and monetize with another simple yet effective app in the ecosystem.
References
https://tapmydata.com/
https://medium.com/@tholder/we-dont-want-your-data-pushing-boundaries-in-data-collection-and-end-to-end-encryption-for-apps-ebd1d5f79df5
![You are the product [RB] (Ep. 147)](https://pbcdn1.podbean.com/imglogo/image-logo/1799802/dsh-cover-2_300x300.jpg)
Sunday Apr 11, 2021
You are the product [RB] (Ep. 147)
Sunday Apr 11, 2021
Sunday Apr 11, 2021
In this episode I am with George Hosu from Cerebralab
and we speak about how dangerous it is not to pay for the services you use, and as a consequence how dangerous it is letting an algorithm decide what you like or not.
Our Sponsors
This episode is supported by Chapman’s Schmid College of Science and Technology, where master’s and PhD students join in cutting-edge research as they prepare to take the next big leap in their professional journey.To learn more about the innovative tools and collaborative approach that distinguish the Chapman program in Computational and Data Sciences, visit chapman.edu/datascience
If building software is your passion, you’ll love ThoughtWorks Technology Podcast. It’s a podcast for techies by techies. Their team of experienced technologists take a deep dive into a tech topic that’s piqued their interest — it could be how machine learning is being used in astrophysics or maybe how to succeed at continuous delivery.
Links
https://cerebralab.com
https://www.eugenewei.com/blog/2019/2/19/status-as-a-service