From 15f8feead5351d3d7b35cc04bdf204f8980c5f8c Mon Sep 17 00:00:00 2001 From: andraghetti Date: Fri, 10 Jan 2025 19:02:17 +0000 Subject: [PATCH] =?UTF-8?q?Deploying=20to=20gh-pages=20from=20@=20theo2021?= =?UTF-8?q?/The-Good-AI-Lab@23ecedbca9d1798b7a4a930c4ab31c1a810e38aa=20?= =?UTF-8?q?=F0=9F=9A=80?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- assets/{index-BNSA-ovO.js => index-De-aW1k1.js} | 2 +- index.html | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) rename assets/{index-BNSA-ovO.js => index-De-aW1k1.js} (98%) diff --git a/assets/index-BNSA-ovO.js b/assets/index-De-aW1k1.js similarity index 98% rename from assets/index-BNSA-ovO.js rename to assets/index-De-aW1k1.js index 4b2a7c5..19f4afa 100644 --- a/assets/index-BNSA-ovO.js +++ b/assets/index-De-aW1k1.js @@ -134,4 +134,4 @@ Error generating stack: `+i.message+` * * This source code is licensed under the ISC license. * See the LICENSE file in the root directory of this source tree. - */const Vh=Le("X",[["path",{d:"M18 6 6 18",key:"1bl5f8"}],["path",{d:"m6 6 12 12",key:"d8bk6v"}]]);function Ss(){const{theme:e,toggleTheme:t}=Th();return S.useEffect(()=>{localStorage.setItem("theme",e),document.documentElement.classList.toggle("dark",e==="dark")},[e]),p.jsx("button",{onClick:t,className:"p-2 rounded-full hover:bg-gray-100 dark:hover:bg-gray-700 transition-colors","aria-label":"Toggle theme",children:e==="dark"?p.jsx(Wh,{className:"h-5 w-5 text-gray-600 dark:text-gray-300"}):p.jsx(Bh,{className:"h-5 w-5 text-gray-600 dark:text-gray-300"})})}function Hh(e){const t=e,n=typeof t.title=="string"&&typeof t.abstract=="string"&&typeof t.image=="string"&&Array.isArray(t.tags)&&t.tags.every(r=>typeof r=="string");return n||console.warn("Invalid research data:",{paper:t,validationErrors:{title:typeof t.title!="string"?"Invalid title type":null,abstract:typeof t.abstract!="string"?"Invalid abstract type":null,image:typeof t.image!="string"?"Invalid image type":null,tags:!Array.isArray(t.tags)||!t.tags.every(r=>typeof r=="string")?"Invalid tags array":null}}),n}function Qh(e){const t=e,n=typeof t.title=="string"&&Array.isArray(t.authors)&&t.authors.every(r=>typeof r=="string")&&typeof t.conference=="string"&&typeof t.year=="number"&&typeof t.abstract=="string"&&typeof t.link=="string"&&typeof t.image=="string";return n||console.warn("Invalid paper data:",{paper:t,validationErrors:{title:typeof t.title!="string"?"Invalid title type":null,authors:!Array.isArray(t.authors)||!t.authors.every(r=>typeof r=="string")?"Invalid authors array":null,conference:typeof t.conference!="string"?"Invalid conference type":null,year:typeof t.year!="number"?"Invalid year type":null,abstract:typeof t.abstract!="string"?"Invalid abstract type":null,link:typeof t.link!="string"?"Invalid link type":null,image:typeof t.image!="string"?"Invalid image type":null}}),n}function Kh(e){const t=e,n=typeof t.name=="string"&&typeof t.role=="string"&&typeof t.image=="string"&&typeof t.bio=="string"&&Array.isArray(t.titles)&&t.titles.every(r=>typeof r=="string")&&typeof t.links=="object"&&(!t.links.scholar||typeof t.links.scholar=="string")&&(!t.links.linkedin||typeof t.links.linkedin=="string");return n||console.warn("Invalid team member data:",{member:t,validationErrors:{name:typeof t.name!="string"?"Invalid name type":null,role:typeof t.role!="string"?"Invalid role type":null,image:typeof t.image!="string"?"Invalid image type":null,bio:typeof t.bio!="string"?"Invalid bio type":null,expertise:!Array.isArray(t.titles)||!t.titles.every(r=>typeof r=="string")?"Invalid titles array":null,links:typeof t.links!="object"?"Invalid links object":null}}),n}const ro="mailto:goodailab@gmail.com?subject=Inquiry%20About%20Joining%20The%20Good%20AI%20Lab&body=Hello%20Good%20AI%20Lab%20Team,%0A%0AI%20would%20like%20to%20learn%20more%20about%20your%20organization%20and%20the%20opportunities%20to%20get%20involved.%0A%0ABest%20regards,%0A[Your%20Name]";function Gh(){const[e,t]=S.useState(!1);return p.jsx("nav",{className:"bg-white dark:bg-gray-800 shadow-sm fixed w-full z-10",children:p.jsxs("div",{className:"max-w-7xl mx-auto px-4 sm:px-6 lg:px-8",children:[p.jsxs("div",{className:"flex justify-between h-16",children:[p.jsxs(Se,{to:"/",className:"flex items-center",children:[p.jsx(Yc,{className:"h-8 w-8 text-green-500"}),p.jsxs("span",{className:"ml-2 text-xl font-semibold dark:text-white",children:["The ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent",children:"Good"})," AI Lab"]})]}),p.jsx("div",{className:"flex items-center md:hidden",children:p.jsx("button",{onClick:()=>t(!e),className:"text-gray-700 dark:text-gray-200 hover:text-green-500",children:e?p.jsx(Vh,{className:"h-6 w-6"}):p.jsx(Uh,{className:"h-6 w-6"})})}),p.jsxs("div",{className:"hidden md:flex items-center space-x-8",children:[p.jsx(Se,{to:"/manifesto",className:"text-gray-700 dark:text-gray-200 hover:text-green-500",children:"Manifesto"}),p.jsx(Se,{to:"/team",className:"text-gray-700 dark:text-gray-200 hover:text-green-500",children:"Team"}),p.jsx(Se,{to:"/research",className:"text-gray-700 dark:text-gray-200 hover:text-green-500",children:"Research"}),p.jsx(Ss,{}),p.jsx(Se,{to:ro,className:"bg-green-500 text-white px-4 py-2 rounded-md hover:bg-green-600 transition-colors",children:"Get Involved"})]})]}),e&&p.jsx("div",{className:"md:hidden",children:p.jsxs("div",{className:"px-2 pt-2 pb-3 space-y-1",children:[p.jsx(Se,{to:"/manifesto",className:"block text-gray-700 dark:text-gray-200 hover:text-green-500 py-2",onClick:()=>t(!1),children:"Manifesto"}),p.jsx(Se,{to:"/team",className:"block text-gray-700 dark:text-gray-200 hover:text-green-500 py-2",onClick:()=>t(!1),children:"Team"}),p.jsx(Se,{to:"/research",className:"block text-gray-700 dark:text-gray-200 hover:text-green-500 py-2",onClick:()=>t(!1),children:"Research"}),p.jsx("div",{className:"py-2",children:p.jsx(Ss,{})}),p.jsx(Se,{to:ro,className:"block bg-green-500 text-white px-4 py-2 rounded-md hover:bg-green-600 transition-colors",children:"Get Involved"})]})})]})})}function Yh(){const e=t=>{t.includes("#")||window.scrollTo({top:0,behavior:"smooth"})};return p.jsx("footer",{className:"bg-gray-900 text-gray-300",children:p.jsxs("div",{className:"max-w-7xl mx-auto px-4 sm:px-6 lg:px-8 py-12",children:[p.jsxs("div",{className:"grid grid-cols-1 md:grid-cols-4 gap-8",children:[p.jsxs("div",{children:[p.jsxs("div",{className:"flex items-center mb-4",children:[p.jsx(Yc,{className:"h-6 w-6 text-green-500"}),p.jsxs("span",{className:"ml-2 text-lg font-semibold text-white",children:["The ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent",children:"Good"})," AI Lab"]})]}),p.jsxs("p",{className:"text-xs",children:["We are an independent AI lab working on foundational"," ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent font-bold",children:"AI research"})," ","and innovation projects for"," ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent font-bold",children:"social good"}),"."]})]}),[{title:"About",links:[{name:"Manifesto",to:"/manifesto"},{name:"Team",to:"/team"}]},{title:"Resources",links:[{name:"Research",to:"/research"},{name:"Community Papers",to:"/research#community"}]},{title:"Get Involved",links:[{name:"Contact us",to:ro}]}].map((t,n)=>p.jsxs("div",{className:"mx-10",children:[p.jsx("h3",{className:"text-white font-semibold mb-4",children:t.title}),p.jsx("ul",{className:"space-y-2",children:t.links.map((r,l)=>p.jsx("li",{children:p.jsx(Se,{to:r.to,className:"hover:text-green-500 transition-colors",onClick:()=>e(r.to),children:r.name})},l))})]},n))]}),p.jsxs("div",{className:"border-t border-gray-800 mt-12 pt-8 text-sm text-center",children:["© ",new Date().getFullYear()," The Good AI Lab. All rights reserved."]})]})})}function Jh(){return p.jsx("div",{className:"min-h-screen flex items-center bg-gradient-to-br from-green-50 to-white dark:from-gray-900 dark:to-gray-800 py-8 sm:py-12 lg:py-8",children:p.jsx("div",{className:"w-full max-w-7xl mx-auto px-4 sm:px-6 lg:px-8 mt-16 sm:mt-0",children:p.jsxs("div",{className:"flex flex-col lg:flex-row lg:items-center lg:gap-12",children:[p.jsxs("div",{className:"flex-1 space-y-6 lg:space-y-16",children:[p.jsx("h1",{className:"text-3xl lg:text-6xl font-bold text-gray-900 dark:text-white mb-4 lg:mb-8 flex flex-wrap items-center justify-center lg:justify-start gap-3 lg:gap-6",children:p.jsxs("span",{className:"text-center lg:text-left",children:["The ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent",children:"Good"})," AI Lab"]})}),p.jsx("div",{className:"prose dark:prose-invert space-y-4 lg:space-y-12",children:p.jsxs("p",{className:"text-xl lg:text-3xl text-gray-600 dark:text-gray-300 leading-relaxed text-center lg:text-left",children:["We are an independent AI lab working on foundational"," ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent font-bold",children:"AI research"})," ","and innovation projects for"," ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent font-bold",children:"social good"}),"."]})}),p.jsx("div",{className:"block lg:hidden w-full my-6",children:p.jsxs("div",{className:"relative max-w-sm mx-auto",children:[p.jsx("div",{className:"absolute -inset-4 rounded-[20%] bg-green-500/20 blur-xl"}),p.jsx("img",{src:"/images/intro.webp",alt:"AI Research",className:"relative rounded-[20%] shadow-2xl"})]})}),p.jsxs("div",{className:"flex flex-col sm:flex-row items-center justify-center lg:justify-start gap-4 lg:gap-8 mt-6 lg:mt-16",children:[p.jsxs(Se,{to:"/manifesto",className:"w-full sm:w-auto bg-green-500 text-white px-6 lg:px-10 py-2.5 lg:py-4 rounded-full hover:bg-green-600 transition-colors flex items-center justify-center text-base lg:text-xl",children:["Learn More",p.jsx(Ah,{className:"ml-2 lg:ml-3 h-5 w-5 lg:h-6 lg:w-6"})]}),p.jsx(Se,{to:"/team",className:"w-full sm:w-auto border-2 border-green-500 text-green-500 px-6 lg:px-10 py-2.5 lg:py-4 rounded-full hover:bg-green-50 dark:hover:bg-green-900/20 transition-colors text-center text-base lg:text-xl",children:"Meet the Team"})]})]}),p.jsx("div",{className:"hidden lg:block lg:w-2/5",children:p.jsxs("div",{className:"relative",children:[p.jsx("div",{className:"absolute -inset-4 rounded-[5%] bg-green-500/20 blur-xl"}),p.jsx("img",{src:"/images/intro.webp",alt:"AI Research",className:"relative rounded-[5%] shadow-2xl"})]})})]})})})}function Xh(){return p.jsx(p.Fragment,{children:p.jsx(Jh,{})})}const Zh={members:[{name:"Pier Luigi Dovesi",role:"AI Scientist @ AMD Silo AI",titles:["Chair of the Board"],image:"/images/team/pier.jpg",bio:"Pier is an AI scientist and researcher with a background in autonomous driving and generative AI, collaborating with universities and research labs worldwide. Pier has a long history of social engagement, actively contributing to initiatives that foster community development and advocate for social change",links:{scholar:"https://scholar.google.com/citations?user=t55rm6EAAAAJ&hl=en",linkedin:"https://www.linkedin.com/in/pierluigidovesi/"}},{name:"Jussi Karlgren",role:"AI Scientist @ AMD Silo AI",titles:["Board Member"],image:"/images/team/jussi.jpg",bio:"Jussi is a computational linguist and AI professor with a PhD in Computational Linguistics from Stockholm University. He specializes in natural language understanding and multilingual language processing, blending academic expertise with industry impact to improve human-machine interaction.",links:{scholar:"https://scholar.google.com/citations?user=o7cz97l31z8J",linkedin:"https://www.linkedin.com/in/jussikarlgren/"}},{name:"Saroosh Shabbir",role:"AI Scientist @ AMD Silo AI",titles:["Board Member","Secretary"],image:"/images/team/saroosh.jpg",bio:"Saroosh holds a PhD in Quantum Optics from KTH. With extensive industry experience, she brings a broad spectrum of expertise spanning deep learning, operations research and quantum computation.",links:{scholar:"https://scholar.google.com/citations?user=qo8efuQAAAAJ",linkedin:"https://www.linkedin.com/in/sarooshsh/"}},{name:"Lorenzo Andraghetti",role:"AI Engineer @ AMD Silo AI",titles:["Board Member","Treasurer"],image:"/images/team/lorenzo.png",bio:"Lorenzo is a senior AI engineer and full-stack software developer with a strong passion for game development. In his most recent engagements, Lorenzo has helped develop large language models for resource scarce Nordic languages.",links:{scholar:"https://scholar.google.com/citations?user=l27rGkIAAAAJ",linkedin:"https://www.linkedin.com/in/andraghetti/"}},{name:"Theodor Panagiotakopoulos",role:"ML Engineer @ King",titles:["Board Member","Auditor"],image:"/images/team/theo.jpeg",bio:"Theo is a machine learning engineer with a strong background in MLOps, combining research insights with practical engineering skills. He is an experienced researcher in domain adaptation, bridging cutting-edge AI advancements with efficient, scalable deployments.",links:{scholar:"https://scholar.google.com/citations?user=868888AAAAAJ",linkedin:"https://www.linkedin.com/in/theopana/"}},{name:"Carlo Rapisarda",role:"AI Engineer @ Bontouch",titles:["Board Member"],image:"/images/team/carlo.jpeg",bio:"Carlo is an experienced AI engineer, software architect, and tech lead with deep expertise in both mobile applications and a range of AI solutions. He has led multiple high-impact projects from design to production, delivering innovative and scalable products across diverse domains.",links:{scholar:"https://scholar.google.com/citations?user=1hsIvRoAAAAJ",linkedin:"https://www.linkedin.com/in/carlorapisarda/"}},{name:"Hedvig Kjellström",role:"Professor @ KTH",titles:["Advisor"],image:"/images/team/hedvig.jpg",bio:"Hedvig is a professor and AI researcher with expertise in computer vision, machine learning, and human behavior modeling. She focuses on developing AI systems that interpret and interact with the physical and social world, bridging theoretical advances with real-world applications.",links:{scholar:"https://scholar.google.com/citations?user=wr3CtKAAAAAJ",linkedin:"https://www.linkedin.com/in/hedvig-kjellstrom/"}},{name:"Matteo Poggi",role:"Assistant Professor @ University of Bologna",titles:["Advisor"],image:"/images/team/matteo.jpeg",bio:"Matteo is an assistant professor and AI researcher at the University of Bologna, specializing in computer vision and deep learning. His work focuses on 3D perception, stereo vision, and scene understanding, with applications ranging from robotics to autonomous systems.",links:{scholar:"https://scholar.google.com/citations?user=bve0VwgAAAAJ",linkedin:"https://www.linkedin.com/in/matteo-poggi-71046a4a/"}},{name:"Riina Mathlin",role:"People Specialist @ AMD Silo AI",titles:["Community Manager"],image:"/images/team/riina.jpeg",bio:"Riina is a people specialist responsible for employee wellbeing, engagement, and lifecycle management. She brings expertise in organisational development and process improvement. She holds MSc in Economics and Business Administration and a BBA in HR Management.",links:{linkedin:"https://www.linkedin.com/in/riinamathlin/"}},{name:"Alexander Aurell",role:"AI Scientist and Manager @ AMD Silo AI",titles:["Researcher"],image:"/images/team/alexander.jpg",bio:"Alexander is an AI scientist and manager with a PhD in mathematics from KTH and postdoctoral work at Princeton, specializing in game theory and modeling for finance and epidemiology. He also brings extensive machine learning expertise for logistics",links:{scholar:"https://scholar.google.com/citations?user=fRoAXJAAAAAJ",linkedin:"https://www.linkedin.com/in/alexander-aurell/"}},{name:"Mattia Segù",role:"PhD Candidate @ ETH & Google",titles:["Researcher"],image:"/images/team/mattia.jpeg",bio:"Mattia is a researcher specializing in computer vision and machine learning. His work focuses on label-efficient multiple object tracking, domain generalization, and uncertainty estimation. Mattia combines deep theoretical expertise with practical applications, contributing to advancements in AI and computer vision.",links:{scholar:"https://scholar.google.com/citations?user=dwX7yWkAAAAJ",linkedin:"https://www.linkedin.com/in/mattia-segu/"}},{name:"Silvio Mazzucco",role:"Computer Vision Researcher @ Huawei",titles:["Researcher"],image:"/images/team/silvio.jpeg",bio:"Silvio is a computer vision researcher at Huawei. Previously at the Computer Vision Lab at ETH Zurich, he specialized in multimodal AI and domain adaptation.",links:{scholar:null,linkedin:"https://www.linkedin.com/in/silvio-mazzucco-7a9672234/"}},{name:"Gianluca Villani",role:"PhD Candidate @ University of Toronto",titles:["Researcher"],image:"/images/team/gianluca.jpeg",bio:"Gianluca is a researcher at the University of Toronto, specializing in the modeling, analysis, and control of biological systems. He combines expertise in AI, mathematics, and control to tackle complex interdisciplinary challenges.",links:{scholar:null,linkedin:"https://www.linkedin.com/in/gianluca-villani-54112b14b/"}},{name:"Linus Härenstam-Nielsen",role:"PhD Candidate @ TU Munich",titles:["Researcher"],image:"/images/team/linus.jpeg",bio:"Linus is a researcher with extensive expertise in computer vision, focusing on learning-based methods, SLAM, and 3D representation. With a strong background in mathematics, robotics, and autonomous systems, he bridges academic research and industry innovation.",links:{scholar:"https://scholar.google.com/citations?user=HWAAVg8AAAAJ",linkedin:"https://www.linkedin.com/in/linus-h%C3%A4renstam-nielsen-41997659/"}},{name:"Leonardo Ravaglia",role:"R&D Researcher @ IMEC",titles:["Researcher"],image:"/images/team/leonardo.jpeg",bio:"Leonardo is a researcher and AI engineer with expertise in autonomous driving and simulation environments. He brings experience in on-device training for ultra-low-power AI accelerators and continual learning, developed during his time at the University of Bologna.",links:{scholar:"https://scholar.google.com/citations?user=hrYEGQIAAAAJ",linkedin:"https://www.linkedin.com/in/leonardo-ravaglia/"}},{name:"Marc Botet Colomer",role:"Deep Learning Engineer @ Univrses",titles:["Researcher"],image:"/images/team/marc.jpeg",bio:"Marc is an AI R&D deep learning engineer with a strong background in autonomous driving. He is an active researcher specializing in domain adaptation and real-time training.",links:{scholar:"https://scholar.google.com/citations?user=-eA8kwwAAAAJ",linkedin:"https://www.linkedin.com/in/marc-botet-colomer/"}},{name:"Carl Persson",role:"AI Engineer @ AMD Silo AI",titles:["Researcher"],image:"/images/team/carl.jpeg",bio:"Carl is an AI engineer specializing in GPU optimization and diffusion models. He is also an researcher, focusing on 3D generative models and domain adaptation.",links:{scholar:null,linkedin:"https://www.linkedin.com/in/carl-persson1/"}},{name:"João Frederico Carvalho",role:"AI Scientist @ Univrses",titles:["Researcher"],image:"/images/team/joao.jpeg",bio:"João holds a PhD in Robotics and Computer Vision from KTH. He is a senior scientist and software developer with extensive experience in autonomous driving, combining deep technical expertise with practical applications in cutting-edge technologies.",links:{scholar:"https://scholar.google.se/citations?user=-GqJRNoAAAAJ",linkedin:"https://www.linkedin.com/in/jo%C3%A3o-frederico-carvalho-27665367/"}}]},qh=Zh.members.filter(Kh);function bh(){return p.jsx("div",{className:"pt-24",children:p.jsxs("div",{className:"max-w-7xl mx-auto px-4 sm:px-6 lg:px-8",children:[p.jsxs("div",{className:"text-center mb-16",children:[p.jsx(Jc,{className:"h-12 w-12 mx-auto text-green-500 mb-4"}),p.jsx("h1",{className:"text-4xl font-bold text-gray-900 dark:text-white mb-4",children:"Team"}),p.jsx("p",{className:"text-xl text-gray-600 dark:text-gray-300 max-w-3xl mx-auto",children:"A collective of scientists, professors, engineers, researchers, and developers coming together from top universities, labs, and companies united by a passion for advancing AI research and driving social impact."})]}),p.jsx("div",{className:"grid md:grid-cols-2 lg:grid-cols-3 gap-8 mb-20",children:qh.map((e,t)=>p.jsxs("div",{className:"bg-white dark:bg-gray-800 rounded-xl shadow-lg overflow-hidden hover:shadow-xl transition-shadow flex flex-col h-full",children:[p.jsxs("div",{className:"relative",children:[p.jsx("img",{className:"w-full h-full object-cover aspect-square",src:e.image,alt:e.name}),p.jsxs("div",{className:"absolute bottom-0 left-0 right-0 bg-gradient-to-t from-black/70 to-transparent p-6",children:[p.jsx("h3",{className:"text-2xl font-bold text-white drop-shadow-[0_2px_4px_rgba(0,0,0,0.8)]",children:e.name}),p.jsx("p",{className:"text-green-500 font-medium drop-shadow-[0_2px_4px_rgba(1.0,1.0,1.0,0.8)]",children:e.role})]})]}),p.jsxs("div",{className:"p-6 flex flex-col flex-grow",children:[p.jsxs("div",{className:"flex-grow space-y-4",children:[p.jsx("div",{className:"flex flex-wrap gap-2",children:e.titles.map((n,r)=>p.jsx("span",{className:"bg-green-50 dark:bg-green-900/30 text-green-700 dark:text-green-300 px-2 py-1 rounded-full text-xs",children:n},r))}),p.jsx("p",{className:"text-gray-600 dark:text-gray-300 line-clamp-10",children:e.bio})]}),p.jsxs("div",{className:"flex gap-4 mt-6 pt-4 border-t border-gray-200 dark:border-gray-700",children:[e.links.scholar&&p.jsxs("a",{href:e.links.scholar,target:"_blank",rel:"noopener noreferrer",className:"flex items-center text-gray-600 dark:text-gray-300 hover:text-green-500 dark:hover:text-green-400",children:[p.jsx(Mh,{className:"h-5 w-5 mr-1"}),"Scholar"]}),e.links.linkedin&&p.jsxs("a",{href:e.links.linkedin,target:"_blank",rel:"noopener noreferrer",className:"flex items-center text-gray-600 dark:text-gray-300 hover:text-green-500 dark:hover:text-green-400",children:[p.jsx(Fh,{className:"h-5 w-5 mr-1"}),"LinkedIn"]})]})]})]},t))})]})})}const em={research:[{title:"Adaptive Agents",abstract:"How can AI agents adapt to ever-changing environments? By seamlessly integrating advanced perception, Q&A, and control strategies—alongside techniques like test-time training, model merging, and on-the-fly knowledge updates—we aim to develop robust, lifelong-learning systems that continuously refine their capabilities in response to new rules, tasks, and scenarios.",image:"/images/research/adaptive-agents.webp",tags:["Agents","Domain Adaptation","Continual Learning"]},{title:"Scene Reconstruction",abstract:"Text-to-image diffusion models excel at generating high-quality visuals, but reconstructing entire scenarios is challenging due to greater spatial complexity. In this project, we propose a pipeline that synthesizes realistic 3D environments from minimal input, combining texturing and semantic preservation with pathfinding and 3D Gaussian Splatting.",image:"/images/research/scene-reconstruction.webp",tags:["Scene Reconstruction","Diffusion","3D Gaussian Splatting"]}]},lo=em.research,io=lo.filter(Hh);io.length!==lo.length&&console.warn(`Filtered out ${lo.length-io.length} invalid papers`);function tm(){return p.jsx("div",{className:"py-16 bg-gray-50 dark:bg-gray-900",children:p.jsx("div",{className:"max-w-7xl mx-auto px-4",children:p.jsxs("div",{className:"grid gap-8 mx-4",children:[p.jsxs("div",{className:"max-w-3xl mx-auto px-4 text-center",children:[p.jsx(Dh,{className:"h-12 w-12 mx-auto text-green-500 dark:text-green-400 mb-6"}),p.jsx("h1",{className:"text-4xl font-bold text-gray-900 dark:text-gray-50 mb-6",children:"Research"}),p.jsxs("p",{className:"text-xl text-gray-600 dark:text-gray-300",children:["We focus on foundational academic AI research.",p.jsx("br",{}),"We publish in open venues, share our code, and push the field forward."]})]}),io.map((e,t)=>p.jsx("div",{className:"bg-white dark:bg-gray-800 rounded-xl shadow-lg overflow-hidden hover:shadow-xl transition-shadow",children:p.jsxs("div",{className:"relative md:flex",children:[p.jsx("div",{className:"md:absolute md:inset-y-0 md:w-64 h-64 md:h-full",children:p.jsx("img",{className:"w-full h-full object-cover",src:e.image,alt:e.title})}),p.jsxs("div",{className:"p-8 md:ml-64",children:[p.jsx("h3",{className:"text-xl font-bold text-gray-900 dark:text-white mb-2",children:e.title}),p.jsx("p",{className:"text-gray-600 dark:text-gray-300 mb-4",children:e.abstract}),p.jsx("div",{className:"flex flex-wrap gap-2 mb-4",children:e.tags.map((n,r)=>p.jsx("span",{className:"bg-green-50 dark:bg-green-900/30 text-green-700 dark:text-green-300 px-3 py-1 rounded-full text-sm",children:n},r))})]})]})},t))]})})})}const nm={papers:[{title:"To Adapt or Not to Adapt? Real-Time Adaptation for Semantic Segmentation",authors:["Marc Botet Colomer","Pier Luigi Dovesi","Theodoros Panagiotakopoulos","Joao Frederico Carvalho","Linus Härenstam-Nielsen","Hossein Azizpour","Hedvig Kjellström","Daniel Cremers","Matteo Poggi"],conference:"ICCV",year:2023,abstract:"The paper explores real-time adaptation strategies for semantic segmentation models, discussing the trade-offs between adaptation and performance in dynamic environments.",link:"https://arxiv.org/abs/2307.15063",image:"/images/papers/iccv2023.png",tags:["Semantic Segmentation","Real-Time Adaptation","Deep Learning"]},{title:"Online Domain Adaptation for Semantic Segmentation in Ever-Changing Conditions",authors:["Theodoros Panagiotakopoulos","Pier Luigi Dovesi","Linus Härenstam-Nielsen","Matteo Poggi"],conference:"ECCV",year:2022,abstract:"This work introduces a framework for online domain adaptation in semantic segmentation, allowing models to adapt to continuously changing environments without offline retraining.",link:"https://arxiv.org/abs/2207.10667",image:"/images/papers/eccv2022.png",tags:["Domain Adaptation","Semantic Segmentation","Online Learning"]},{title:"Real-Time Semantic Stereo Matching",authors:["Pier Luigi Dovesi","Matteo Poggi","Lorenzo Andraghetti","Miquel Martí","Hedvig Kjellström","Alessandro Pieropan","Stefano Mattoccia"],conference:"ICRA",year:2020,abstract:"This paper presents a compact and lightweight architecture for real-time semantic stereo matching, enabling efficient inference on embedded devices with minimal accuracy loss.",link:"https://arxiv.org/abs/1910.00541",image:"/images/papers/icra2020.png",tags:["Semantic Stereo Matching","Real-Time Processing","Deep Learning"]},{title:"Enhancing Self-Supervised Monocular Depth Estimation with Traditional Visual Odometry",authors:["Lorenzo Andraghetti","Panteleimon Myriokefalitakis","Pier Luigi Dovesi","Belén Luque","Matteo Poggi","Alessandro Pieropan","Stefano Mattoccia"],conference:"3DV",year:2019,abstract:"The authors propose a method to improve self-supervised monocular depth estimation by integrating traditional visual odometry techniques, achieving better accuracy in depth prediction.",link:"https://arxiv.org/abs/1908.03127",image:"/images/papers/3dv2019.png",tags:["Monocular Depth Estimation","Visual Odometry","Self-Supervised Learning"]}]},rm=nm.papers.filter(Qh);function lm(){const e=fr();return S.useEffect(()=>{if(e.hash==="#community"){const t=document.getElementById("community");t&&t.scrollIntoView({behavior:"smooth"})}},[e]),p.jsx("div",{className:"py-16 bg-white dark:bg-gray-800",id:"community",children:p.jsxs("div",{className:"max-w-7xl mx-auto px-4",children:[p.jsxs("div",{className:"flex items-center justify-center gap-3 mb-12",children:[p.jsx(Jc,{className:"h-8 w-8 text-green-500"}),p.jsx("h2",{className:"text-3xl font-bold text-center text-gray-900 dark:text-white",children:"Community Papers"})]}),p.jsx("div",{className:"grid md:grid-cols-2 gap-8",children:rm.map((t,n)=>p.jsxs("div",{className:"bg-gray-50 dark:bg-gray-700 rounded-xl shadow-lg overflow-hidden",children:[p.jsx("img",{className:"w-full h-48 object-cover",src:t.image,alt:t.title}),p.jsxs("div",{className:"p-6",children:[p.jsxs("p",{className:"text-sm text-green-600 dark:text-green-400 mb-2",children:[t.conference," • ",t.year]}),p.jsx("h3",{className:"text-xl font-bold text-gray-900 dark:text-white mb-2",children:t.title}),p.jsx("p",{className:"text-sm text-gray-600 dark:text-gray-300 mb-4",children:t.authors.join(", ")}),p.jsx("p",{className:"text-gray-600 dark:text-gray-300 mb-4",children:t.abstract}),p.jsxs("a",{href:t.link,className:"inline-flex items-center text-green-600 dark:text-green-400 hover:text-green-700 dark:hover:text-green-300",children:[p.jsx(Oh,{className:"h-4 w-4 mr-2"}),"View Publication"]})]})]},n))})]})})}function im(){return p.jsxs("div",{className:"pt-16",children:[p.jsx(tm,{}),p.jsx(lm,{})]})}function om(){return p.jsxs("div",{className:"pt-28 lg:pt-24 bg-white dark:bg-gray-900",children:[p.jsx("div",{className:"max-w-4xl mx-auto px-4 sm:px-6 lg:px-8",children:p.jsxs("div",{className:"text-center mb-8 lg:mb-12",children:[p.jsx($h,{className:"h-10 w-10 lg:h-12 lg:w-12 mx-auto text-green-600 dark:text-green-500 mb-4"}),p.jsx("h1",{className:"text-3xl lg:text-4xl font-bold text-gray-900 dark:text-white mb-4",children:"Our Manifesto"})]})}),p.jsxs("div",{className:"prose prose-base lg:prose-lg mx-auto px-4 lg:px-0 mb-16 lg:mb-20 dark:prose-invert prose-headings:text-gray-900 dark:prose-headings:text-white prose-p:text-gray-600 dark:prose-p:text-gray-300 prose-li:text-gray-600 dark:prose-li:text-gray-300 prose-h3:text-green-600 dark:prose-h3:text-green-500 max-w-4xl lg:max-w-8xl",children:[p.jsx("h3",{className:"text-xl lg:text-2xl mt-8 first:mt-0",children:"Who We Are"}),p.jsx("p",{className:"text-base lg:text-lg",children:"We are a non-profit AI research collective—scientists, professors, engineers, researchers, and developers—coming together from top universities, labs, and companies. We're here because we're curious about what AI can be and should be, and we feel a responsibility to make it count for everyone."}),p.jsx("h3",{className:"text-xl lg:text-2xl mt-8",children:"What We Believe"}),p.jsx("p",{className:"text-base lg:text-lg",children:"AI is shaping the future. It's changing every aspect of our society, including how we think about ourselves and our potential. The question is not just what AI can do, but what it should do. We believe that innovation without a guiding purpose is hollow. Real progress means more than building better tools; it means building them for everyone, especially those too often overlooked or forgotten. We want AI that doesn't just float at the top, serving the privileged, but flows downward and outward, reaching people and places long left behind."}),p.jsx("h3",{className:"text-xl lg:text-2xl mt-8",children:"What We Do"}),p.jsxs("ul",{className:"text-base lg:text-lg space-y-4",children:[p.jsxs("li",{children:[p.jsx("strong",{children:"AI Lab:"})," We focus on foundational academic AI research. We publish in open venues, share our code, and push the field forward."]}),p.jsxs("li",{children:[p.jsx("strong",{children:"Good AI:"})," We call out biases, support frameworks that keep AI honest, and build fairness, transparency, and accountability into every model we touch. We fight for meaningful, responsible applications that truly matter."]})]}),p.jsx("h3",{className:"text-xl lg:text-2xl mt-8",children:"How We Work"}),p.jsxs("ul",{className:"text-base lg:text-lg space-y-4",children:[p.jsxs("li",{children:[p.jsx("strong",{children:"Collaborations:"})," We team up with universities, research institutes, and R&D labs to tackle projects that have real impact."]}),p.jsxs("li",{children:[p.jsx("strong",{children:"Openness and Community:"})," We share what we find, learn, and create at conferences, workshops, and public events, aiming to spark honest conversation and inspire both research and action."]}),p.jsxs("li",{children:[p.jsx("strong",{children:"Partnerships:"})," Hand-in-hand with charitable non-profit organisations, we promote and develop AI applications that drive tangible improvements—education, healthcare, humanitarian aid, environmental protection, and more."]})]}),p.jsx("h3",{className:"text-xl lg:text-2xl mt-8",children:"Our Vision"}),p.jsx("p",{className:"text-base lg:text-lg",children:"We see AI as a force that reveals new paths, includes more voices, and helps reshape society for the better. If we are to contribute with a verse to this powerful play, let it be driven by curiosity and guided by responsibility."})]})]})}function am(){return p.jsx(Rh,{children:p.jsx(Nh,{children:p.jsxs("div",{className:"min-h-screen bg-white dark:bg-gray-900",children:[p.jsx(Gh,{}),p.jsxs(yh,{children:[p.jsx(Rn,{path:"/",element:p.jsx(Xh,{})}),p.jsx(Rn,{path:"/team",element:p.jsx(bh,{})}),p.jsx(Rn,{path:"/research",element:p.jsx(im,{})}),p.jsx(Rn,{path:"/manifesto",element:p.jsx(om,{})})]}),p.jsx(Yh,{})]})})})}Mc(document.getElementById("root")).render(p.jsx(S.StrictMode,{children:p.jsx(am,{})})); + */const Vh=Le("X",[["path",{d:"M18 6 6 18",key:"1bl5f8"}],["path",{d:"m6 6 12 12",key:"d8bk6v"}]]);function Ss(){const{theme:e,toggleTheme:t}=Th();return S.useEffect(()=>{localStorage.setItem("theme",e),document.documentElement.classList.toggle("dark",e==="dark")},[e]),p.jsx("button",{onClick:t,className:"p-2 rounded-full hover:bg-gray-100 dark:hover:bg-gray-700 transition-colors","aria-label":"Toggle theme",children:e==="dark"?p.jsx(Wh,{className:"h-5 w-5 text-gray-600 dark:text-gray-300"}):p.jsx(Bh,{className:"h-5 w-5 text-gray-600 dark:text-gray-300"})})}function Hh(e){const t=e,n=typeof t.title=="string"&&typeof t.abstract=="string"&&typeof t.image=="string"&&Array.isArray(t.tags)&&t.tags.every(r=>typeof r=="string");return n||console.warn("Invalid research data:",{paper:t,validationErrors:{title:typeof t.title!="string"?"Invalid title type":null,abstract:typeof t.abstract!="string"?"Invalid abstract type":null,image:typeof t.image!="string"?"Invalid image type":null,tags:!Array.isArray(t.tags)||!t.tags.every(r=>typeof r=="string")?"Invalid tags array":null}}),n}function Qh(e){const t=e,n=typeof t.title=="string"&&Array.isArray(t.authors)&&t.authors.every(r=>typeof r=="string")&&typeof t.conference=="string"&&typeof t.year=="number"&&typeof t.abstract=="string"&&typeof t.link=="string"&&typeof t.image=="string";return n||console.warn("Invalid paper data:",{paper:t,validationErrors:{title:typeof t.title!="string"?"Invalid title type":null,authors:!Array.isArray(t.authors)||!t.authors.every(r=>typeof r=="string")?"Invalid authors array":null,conference:typeof t.conference!="string"?"Invalid conference type":null,year:typeof t.year!="number"?"Invalid year type":null,abstract:typeof t.abstract!="string"?"Invalid abstract type":null,link:typeof t.link!="string"?"Invalid link type":null,image:typeof t.image!="string"?"Invalid image type":null}}),n}function Kh(e){const t=e,n=typeof t.name=="string"&&typeof t.role=="string"&&typeof t.image=="string"&&typeof t.bio=="string"&&Array.isArray(t.titles)&&t.titles.every(r=>typeof r=="string")&&typeof t.links=="object"&&(!t.links.scholar||typeof t.links.scholar=="string")&&(!t.links.linkedin||typeof t.links.linkedin=="string");return n||console.warn("Invalid team member data:",{member:t,validationErrors:{name:typeof t.name!="string"?"Invalid name type":null,role:typeof t.role!="string"?"Invalid role type":null,image:typeof t.image!="string"?"Invalid image type":null,bio:typeof t.bio!="string"?"Invalid bio type":null,expertise:!Array.isArray(t.titles)||!t.titles.every(r=>typeof r=="string")?"Invalid titles array":null,links:typeof t.links!="object"?"Invalid links object":null}}),n}const ro="mailto:goodailab@gmail.com?subject=Inquiry%20About%20Joining%20The%20Good%20AI%20Lab&body=Hello%20Good%20AI%20Lab%20Team,%0A%0AI%20would%20like%20to%20learn%20more%20about%20your%20organization%20and%20the%20opportunities%20to%20get%20involved.%0A%0ABest%20regards,%0A[Your%20Name]";function Gh(){const[e,t]=S.useState(!1);return p.jsx("nav",{className:"bg-white dark:bg-gray-800 shadow-sm fixed w-full z-10",children:p.jsxs("div",{className:"max-w-7xl mx-auto px-4 sm:px-6 lg:px-8",children:[p.jsxs("div",{className:"flex justify-between h-16",children:[p.jsxs(Se,{to:"/",className:"flex items-center",children:[p.jsx(Yc,{className:"h-8 w-8 text-green-500"}),p.jsxs("span",{className:"ml-2 text-xl font-semibold dark:text-white",children:["The ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent",children:"Good"})," AI Lab"]})]}),p.jsx("div",{className:"flex items-center md:hidden",children:p.jsx("button",{onClick:()=>t(!e),className:"text-gray-700 dark:text-gray-200 hover:text-green-500",children:e?p.jsx(Vh,{className:"h-6 w-6"}):p.jsx(Uh,{className:"h-6 w-6"})})}),p.jsxs("div",{className:"hidden md:flex items-center space-x-8",children:[p.jsx(Se,{to:"/manifesto",className:"text-gray-700 dark:text-gray-200 hover:text-green-500",children:"Manifesto"}),p.jsx(Se,{to:"/team",className:"text-gray-700 dark:text-gray-200 hover:text-green-500",children:"Team"}),p.jsx(Se,{to:"/research",className:"text-gray-700 dark:text-gray-200 hover:text-green-500",children:"Research"}),p.jsx(Ss,{}),p.jsx(Se,{to:ro,className:"bg-green-500 text-white px-4 py-2 rounded-md hover:bg-green-600 transition-colors",children:"Get Involved"})]})]}),e&&p.jsx("div",{className:"md:hidden",children:p.jsxs("div",{className:"px-2 pt-2 pb-3 space-y-1",children:[p.jsx(Se,{to:"/manifesto",className:"block text-gray-700 dark:text-gray-200 hover:text-green-500 py-2",onClick:()=>t(!1),children:"Manifesto"}),p.jsx(Se,{to:"/team",className:"block text-gray-700 dark:text-gray-200 hover:text-green-500 py-2",onClick:()=>t(!1),children:"Team"}),p.jsx(Se,{to:"/research",className:"block text-gray-700 dark:text-gray-200 hover:text-green-500 py-2",onClick:()=>t(!1),children:"Research"}),p.jsx("div",{className:"py-2",children:p.jsx(Ss,{})}),p.jsx(Se,{to:ro,className:"block bg-green-500 text-white px-4 py-2 rounded-md hover:bg-green-600 transition-colors",children:"Get Involved"})]})})]})})}function Yh(){const e=t=>{t.includes("#")||window.scrollTo({top:0,behavior:"smooth"})};return p.jsx("footer",{className:"bg-gray-900 text-gray-300",children:p.jsxs("div",{className:"max-w-7xl mx-auto px-4 sm:px-6 lg:px-8 py-12",children:[p.jsxs("div",{className:"grid grid-cols-1 md:grid-cols-4 gap-8",children:[p.jsxs("div",{children:[p.jsxs("div",{className:"flex items-center mb-4",children:[p.jsx(Yc,{className:"h-6 w-6 text-green-500"}),p.jsxs("span",{className:"ml-2 text-lg font-semibold text-white",children:["The ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent",children:"Good"})," AI Lab"]})]}),p.jsxs("p",{className:"text-xs",children:["We are an independent AI lab working on foundational"," ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent font-bold",children:"AI research"})," ","and innovation projects for"," ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent font-bold",children:"social good"}),"."]})]}),[{title:"About",links:[{name:"Manifesto",to:"/manifesto"},{name:"Team",to:"/team"}]},{title:"Resources",links:[{name:"Research",to:"/research"},{name:"Community Papers",to:"/research#community"}]},{title:"Get Involved",links:[{name:"Contact us",to:ro}]}].map((t,n)=>p.jsxs("div",{className:"mx-10",children:[p.jsx("h3",{className:"text-white font-semibold mb-4",children:t.title}),p.jsx("ul",{className:"space-y-2",children:t.links.map((r,l)=>p.jsx("li",{children:p.jsx(Se,{to:r.to,className:"hover:text-green-500 transition-colors",onClick:()=>e(r.to),children:r.name})},l))})]},n))]}),p.jsxs("div",{className:"border-t border-gray-800 mt-12 pt-8 text-sm text-center",children:["© ",new Date().getFullYear()," The Good AI Lab. All rights reserved."]})]})})}function Jh(){return p.jsx("div",{className:"min-h-screen flex items-center bg-gradient-to-br from-green-50 to-white dark:from-gray-900 dark:to-gray-800 py-8 sm:py-12 lg:py-8",children:p.jsx("div",{className:"w-full max-w-7xl mx-auto px-4 sm:px-6 lg:px-8 mt-16 sm:mt-0",children:p.jsxs("div",{className:"flex flex-col lg:flex-row lg:items-center lg:gap-12",children:[p.jsxs("div",{className:"flex-1 space-y-6 lg:space-y-16",children:[p.jsx("h1",{className:"text-3xl lg:text-6xl font-bold text-gray-900 dark:text-white mb-4 lg:mb-8 flex flex-wrap items-center justify-center lg:justify-start gap-3 lg:gap-6",children:p.jsxs("span",{className:"text-center lg:text-left",children:["The ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent",children:"Good"})," AI Lab"]})}),p.jsx("div",{className:"prose dark:prose-invert space-y-4 lg:space-y-12",children:p.jsxs("p",{className:"text-xl lg:text-3xl text-gray-600 dark:text-gray-300 leading-relaxed text-center lg:text-left",children:["We are an independent AI lab working on foundational"," ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent font-bold",children:"AI research"})," ","and innovation projects for"," ",p.jsx("span",{className:"bg-gradient-to-r from-green-400 to-green-600 bg-clip-text text-transparent font-bold",children:"social good"}),"."]})}),p.jsx("div",{className:"block lg:hidden w-full my-6",children:p.jsxs("div",{className:"relative max-w-sm mx-auto",children:[p.jsx("div",{className:"absolute -inset-4 rounded-[20%] bg-green-500/20 blur-xl"}),p.jsx("img",{src:"/images/intro.webp",alt:"AI Research",className:"relative rounded-[20%] shadow-2xl"})]})}),p.jsxs("div",{className:"flex flex-col sm:flex-row items-center justify-center lg:justify-start gap-4 lg:gap-8 mt-6 lg:mt-16",children:[p.jsxs(Se,{to:"/manifesto",className:"w-full sm:w-auto bg-green-500 text-white px-6 lg:px-10 py-2.5 lg:py-4 rounded-full hover:bg-green-600 transition-colors flex items-center justify-center text-base lg:text-xl",children:["Learn More",p.jsx(Ah,{className:"ml-2 lg:ml-3 h-5 w-5 lg:h-6 lg:w-6"})]}),p.jsx(Se,{to:"/team",className:"w-full sm:w-auto border-2 border-green-500 text-green-500 px-6 lg:px-10 py-2.5 lg:py-4 rounded-full hover:bg-green-50 dark:hover:bg-green-900/20 transition-colors text-center text-base lg:text-xl",children:"Meet the Team"})]})]}),p.jsx("div",{className:"hidden lg:block lg:w-2/5",children:p.jsxs("div",{className:"relative",children:[p.jsx("div",{className:"absolute -inset-4 rounded-[5%] bg-green-500/20 blur-xl"}),p.jsx("img",{src:"/images/intro.webp",alt:"AI Research",className:"relative rounded-[5%] shadow-2xl"})]})})]})})})}function Xh(){return p.jsx(p.Fragment,{children:p.jsx(Jh,{})})}const Zh={members:[{name:"Pier Luigi Dovesi",role:"AI Scientist @ AMD Silo AI",titles:["Chair of the Board"],image:"/images/team/pier.jpg",bio:"Pier is an AI scientist and researcher with a background in autonomous driving and generative AI, collaborating with universities and research labs worldwide. Pier has a long history of social engagement, actively contributing to initiatives that foster community development and advocate for social change",links:{scholar:"https://scholar.google.com/citations?user=t55rm6EAAAAJ&hl=en",linkedin:"https://www.linkedin.com/in/pierluigidovesi/"}},{name:"Jussi Karlgren",role:"AI Scientist @ AMD Silo AI",titles:["Board Member"],image:"/images/team/jussi.jpg",bio:"Jussi is a computational linguist and AI professor with a PhD in Computational Linguistics from Stockholm University. He specializes in natural language understanding and multilingual language processing, blending academic expertise with industry impact to improve human-machine interaction.",links:{scholar:"https://scholar.google.com/citations?user=o7cz97l31z8J",linkedin:"https://www.linkedin.com/in/jussikarlgren/"}},{name:"Saroosh Shabbir",role:"AI Scientist @ AMD Silo AI",titles:["Board Member","Secretary"],image:"/images/team/saroosh.jpg",bio:"Saroosh holds a PhD in Quantum Optics from KTH. With extensive industry experience, she brings a broad spectrum of expertise spanning deep learning, operations research and quantum computation.",links:{scholar:"https://scholar.google.com/citations?user=qo8efuQAAAAJ",linkedin:"https://www.linkedin.com/in/sarooshsh/"}},{name:"Lorenzo Andraghetti",role:"AI Engineer @ AMD Silo AI",titles:["Board Member","Treasurer"],image:"/images/team/lorenzo.png",bio:"Lorenzo is a senior AI engineer and full-stack software developer with a strong passion for game development. In his most recent engagements, Lorenzo has helped develop large language models for resource scarce Nordic languages.",links:{scholar:"https://scholar.google.com/citations?user=l27rGkIAAAAJ",linkedin:"https://www.linkedin.com/in/andraghetti/"}},{name:"Theodor Panagiotakopoulos",role:"ML Engineer @ King",titles:["Board Member","Auditor"],image:"/images/team/theo.jpeg",bio:"Theo is a machine learning engineer with a strong background in MLOps, combining research insights with practical engineering skills. He is an experienced researcher in domain adaptation, bridging cutting-edge AI advancements with efficient, scalable deployments.",links:{scholar:"https://scholar.google.com/citations?user=868888AAAAAJ",linkedin:"https://www.linkedin.com/in/theopana/"}},{name:"Carlo Rapisarda",role:"AI Engineer @ Bontouch",titles:["Board Member"],image:"/images/team/carlo.jpeg",bio:"Carlo is an experienced AI engineer, software architect, and tech lead with deep expertise in both mobile applications and a range of AI solutions. He has led multiple high-impact projects from design to production, delivering innovative and scalable products across diverse domains.",links:{scholar:"https://scholar.google.com/citations?user=1hsIvRoAAAAJ",linkedin:"https://www.linkedin.com/in/carlorapisarda/"}},{name:"Hedvig Kjellström",role:"Professor @ KTH",titles:["Advisor"],image:"/images/team/hedvig.jpg",bio:"Hedvig is a professor and AI researcher with expertise in computer vision, machine learning, and human behavior modeling. She focuses on developing AI systems that interpret and interact with the physical and social world, bridging theoretical advances with real-world applications.",links:{scholar:"https://scholar.google.com/citations?user=wr3CtKAAAAAJ",linkedin:"https://www.linkedin.com/in/hedvig-kjellstrom/"}},{name:"Matteo Poggi",role:"Assistant Professor @ University of Bologna",titles:["Advisor"],image:"/images/team/matteo.jpeg",bio:"Matteo is an assistant professor and AI researcher at the University of Bologna, specializing in computer vision and deep learning. His work focuses on 3D perception, stereo vision, and scene understanding, with applications ranging from robotics to autonomous systems.",links:{scholar:"https://scholar.google.com/citations?user=bve0VwgAAAAJ",linkedin:"https://www.linkedin.com/in/matteo-poggi-71046a4a/"}},{name:"Riina Mathlin",role:"People Specialist @ AMD Silo AI",titles:["Community Manager"],image:"/images/team/riina.jpeg",bio:"Riina is a people specialist responsible for employee wellbeing, engagement, and lifecycle management. She brings expertise in organisational development and process improvement. She holds MSc in Economics and Business Administration and a BBA in HR Management.",links:{linkedin:"https://www.linkedin.com/in/riinamathlin/"}},{name:"Alexander Aurell",role:"AI Scientist and Manager @ AMD Silo AI",titles:["Researcher"],image:"/images/team/alexander.jpg",bio:"Alexander is an AI scientist and manager with a PhD in mathematics from KTH and postdoctoral work at Princeton, specializing in game theory and modeling for finance and epidemiology. He also brings extensive machine learning expertise for logistics",links:{scholar:"https://scholar.google.com/citations?user=fRoAXJAAAAAJ",linkedin:"https://www.linkedin.com/in/alexander-aurell/"}},{name:"Mattia Segù",role:"PhD Candidate @ ETH & Google",titles:["Researcher"],image:"/images/team/mattia.jpeg",bio:"Mattia is a researcher specializing in computer vision and machine learning. His work focuses on label-efficient multiple object tracking, domain generalization, and uncertainty estimation. Mattia combines deep theoretical expertise with practical applications, contributing to advancements in AI and computer vision.",links:{scholar:"https://scholar.google.com/citations?user=dwX7yWkAAAAJ",linkedin:"https://www.linkedin.com/in/mattia-segu/"}},{name:"Silvio Mazzucco",role:"Computer Vision Researcher @ Huawei",titles:["Researcher"],image:"/images/team/silvio.jpeg",bio:"Silvio is a computer vision researcher at Huawei. Previously at the Computer Vision Lab at ETH Zurich, he specialized in multimodal AI and domain adaptation.",links:{scholar:null,linkedin:"https://www.linkedin.com/in/silvio-mazzucco-7a9672234/"}},{name:"Gianluca Villani",role:"PhD Candidate @ University of Toronto",titles:["Researcher"],image:"/images/team/gianluca.jpeg",bio:"Gianluca is a researcher at the University of Toronto, specializing in the modeling, analysis, and control of biological systems. He combines expertise in AI, mathematics, and control to tackle complex interdisciplinary challenges.",links:{scholar:null,linkedin:"https://www.linkedin.com/in/gianluca-villani-54112b14b/"}},{name:"Linus Härenstam-Nielsen",role:"PhD Candidate @ TU Munich",titles:["Researcher"],image:"/images/team/linus.jpeg",bio:"Linus is a researcher with extensive expertise in computer vision, focusing on learning-based methods, SLAM, and 3D representation. With a strong background in mathematics, robotics, and autonomous systems, he bridges academic research and industry innovation.",links:{scholar:"https://scholar.google.com/citations?user=HWAAVg8AAAAJ",linkedin:"https://www.linkedin.com/in/linus-h%C3%A4renstam-nielsen-41997659/"}},{name:"Leonardo Ravaglia",role:"R&D Researcher @ IMEC",titles:["Researcher"],image:"/images/team/leonardo.jpeg",bio:"Leonardo is a researcher and AI engineer with expertise in autonomous driving and simulation environments. He brings experience in on-device training for ultra-low-power AI accelerators and continual learning, developed during his time at the University of Bologna.",links:{scholar:"https://scholar.google.com/citations?user=hrYEGQIAAAAJ",linkedin:"https://www.linkedin.com/in/leonardo-ravaglia/"}},{name:"Marc Botet Colomer",role:"Deep Learning Engineer @ Univrses",titles:["Researcher"],image:"/images/team/marc.jpeg",bio:"Marc is an AI R&D deep learning engineer with a strong background in autonomous driving. He is an active researcher specializing in domain adaptation and real-time training.",links:{scholar:"https://scholar.google.com/citations?user=-eA8kwwAAAAJ",linkedin:"https://www.linkedin.com/in/marc-botet-colomer/"}},{name:"Carl Persson",role:"AI Engineer @ AMD Silo AI",titles:["Researcher"],image:"/images/team/carl.jpeg",bio:"Carl is an AI engineer specializing in GPU optimization and diffusion models. He is also an researcher, focusing on 3D generative models and domain adaptation.",links:{scholar:null,linkedin:"https://www.linkedin.com/in/carl-persson1/"}},{name:"João Frederico Carvalho",role:"AI Scientist @ Univrses",titles:["Researcher"],image:"/images/team/joao.jpeg",bio:"João holds a PhD in Robotics and Computer Vision from KTH. He is a senior scientist and software developer with extensive experience in autonomous driving, combining deep technical expertise with practical applications in cutting-edge technologies.",links:{scholar:"https://scholar.google.se/citations?user=-GqJRNoAAAAJ",linkedin:"https://www.linkedin.com/in/jo%C3%A3o-frederico-carvalho-27665367/"}}]},qh=Zh.members.filter(Kh);function bh(){return p.jsx("div",{className:"pt-24",children:p.jsxs("div",{className:"max-w-7xl mx-auto px-4 sm:px-6 lg:px-8",children:[p.jsxs("div",{className:"text-center mb-16",children:[p.jsx(Jc,{className:"h-12 w-12 mx-auto text-green-500 mb-4"}),p.jsx("h1",{className:"text-4xl font-bold text-gray-900 dark:text-white mb-4",children:"Team"}),p.jsx("p",{className:"text-xl text-gray-600 dark:text-gray-300 max-w-3xl mx-auto",children:"A collective of scientists, professors, engineers, researchers, and developers coming together from top universities, labs, and companies united by a passion for advancing AI research and driving social impact."})]}),p.jsx("div",{className:"grid md:grid-cols-2 lg:grid-cols-3 gap-8 mb-20",children:qh.map((e,t)=>p.jsxs("div",{className:"bg-white dark:bg-gray-800 rounded-xl shadow-lg overflow-hidden hover:shadow-xl transition-shadow flex flex-col h-full",children:[p.jsxs("div",{className:"relative",children:[p.jsx("img",{className:"w-full h-full object-cover aspect-square",src:e.image,alt:e.name}),p.jsxs("div",{className:"absolute bottom-0 left-0 right-0 bg-gradient-to-t from-black/70 to-transparent p-6",children:[p.jsx("h3",{className:"text-2xl font-bold text-white drop-shadow-[0_2px_4px_rgba(0,0,0,0.8)]",children:e.name}),p.jsx("p",{className:"text-green-500 font-medium drop-shadow-[0_2px_4px_rgba(1.0,1.0,1.0,0.8)]",children:e.role})]})]}),p.jsxs("div",{className:"p-6 flex flex-col flex-grow",children:[p.jsxs("div",{className:"flex-grow space-y-4",children:[p.jsx("div",{className:"flex flex-wrap gap-2",children:e.titles.map((n,r)=>p.jsx("span",{className:"bg-green-50 dark:bg-green-900/30 text-green-700 dark:text-green-300 px-2 py-1 rounded-full text-xs",children:n},r))}),p.jsx("p",{className:"text-gray-600 dark:text-gray-300 line-clamp-10",children:e.bio})]}),p.jsxs("div",{className:"flex gap-4 mt-6 pt-4 border-t border-gray-200 dark:border-gray-700",children:[e.links.scholar&&p.jsxs("a",{href:e.links.scholar,target:"_blank",rel:"noopener noreferrer",className:"flex items-center text-gray-600 dark:text-gray-300 hover:text-green-500 dark:hover:text-green-400",children:[p.jsx(Mh,{className:"h-5 w-5 mr-1"}),"Scholar"]}),e.links.linkedin&&p.jsxs("a",{href:e.links.linkedin,target:"_blank",rel:"noopener noreferrer",className:"flex items-center text-gray-600 dark:text-gray-300 hover:text-green-500 dark:hover:text-green-400",children:[p.jsx(Fh,{className:"h-5 w-5 mr-1"}),"LinkedIn"]})]})]})]},t))})]})})}const em={research:[{title:"Adaptive Agents",abstract:"How can AI agents adapt to ever-changing environments? By seamlessly integrating advanced perception, Q&A, and control strategies—alongside techniques like test-time training, model merging, and on-the-fly knowledge updates—we aim to develop robust, lifelong-learning systems that continuously refine their capabilities in response to new rules, tasks, and scenarios.",image:"/images/research/adaptive-agents.webp",tags:["Agents","Domain Adaptation","Continual Learning"]},{title:"Scene Reconstruction",abstract:"Text-to-image diffusion models excel at generating high-quality visuals, but reconstructing entire scenarios is challenging due to greater spatial complexity. In this project, we propose a pipeline that synthesizes realistic 3D environments from minimal input, combining texturing and semantic preservation with pathfinding and 3D Gaussian Splatting.",image:"/images/research/scene-reconstruction.webp",tags:["Scene Reconstruction","Diffusion","3D Gaussian Splatting"]}]},lo=em.research,io=lo.filter(Hh);io.length!==lo.length&&console.warn(`Filtered out ${lo.length-io.length} invalid papers`);function tm(){return p.jsx("div",{className:"py-16 bg-gray-50 dark:bg-gray-900",children:p.jsx("div",{className:"max-w-7xl mx-auto px-4",children:p.jsxs("div",{className:"grid gap-8 mx-4",children:[p.jsxs("div",{className:"max-w-3xl mx-auto px-4 text-center",children:[p.jsx(Dh,{className:"h-12 w-12 mx-auto text-green-500 dark:text-green-400 mb-6"}),p.jsx("h1",{className:"text-4xl font-bold text-gray-900 dark:text-gray-50 mb-6",children:"Research"}),p.jsxs("p",{className:"text-xl text-gray-600 dark:text-gray-300",children:["We focus on foundational academic AI research.",p.jsx("br",{}),"We publish in open venues, share our code, and push the field forward."]})]}),io.map((e,t)=>p.jsx("div",{className:"bg-white dark:bg-gray-800 rounded-xl shadow-lg overflow-hidden hover:shadow-xl transition-shadow",children:p.jsxs("div",{className:"relative md:flex",children:[p.jsx("div",{className:"md:absolute md:inset-y-0 md:w-64 h-64 md:h-full",children:p.jsx("img",{className:"w-full h-full object-cover",src:e.image,alt:e.title})}),p.jsxs("div",{className:"p-8 md:ml-64",children:[p.jsx("h3",{className:"text-xl font-bold text-gray-900 dark:text-white mb-2",children:e.title}),p.jsx("p",{className:"text-gray-600 dark:text-gray-300 mb-4",children:e.abstract}),p.jsx("div",{className:"flex flex-wrap gap-2 mb-4",children:e.tags.map((n,r)=>p.jsx("span",{className:"bg-green-50 dark:bg-green-900/30 text-green-700 dark:text-green-300 px-3 py-1 rounded-full text-sm",children:n},r))})]})]})},t))]})})})}const nm={papers:[{title:"To Adapt or Not to Adapt? Real-Time Adaptation for Semantic Segmentation",authors:["Marc Botet Colomer","Pier Luigi Dovesi","Theodoros Panagiotakopoulos","Joao Frederico Carvalho","Linus Härenstam-Nielsen","Hossein Azizpour","Hedvig Kjellström","Daniel Cremers","Matteo Poggi"],conference:"ICCV",year:2023,abstract:"The paper explores real-time adaptation strategies for semantic segmentation models, discussing the trade-offs between adaptation and performance in dynamic environments.",link:"https://arxiv.org/abs/2307.15063",image:"/images/papers/iccv2023.png",tags:["Semantic Segmentation","Real-Time Adaptation","Deep Learning"]},{title:"Online Domain Adaptation for Semantic Segmentation in Ever-Changing Conditions",authors:["Theodoros Panagiotakopoulos","Pier Luigi Dovesi","Linus Härenstam-Nielsen","Matteo Poggi"],conference:"ECCV",year:2022,abstract:"This work introduces a framework for online domain adaptation in semantic segmentation, allowing models to adapt to continuously changing environments without offline retraining.",link:"https://arxiv.org/abs/2207.10667",image:"/images/papers/eccv2022.png",tags:["Domain Adaptation","Semantic Segmentation","Online Learning"]},{title:"Real-Time Semantic Stereo Matching",authors:["Pier Luigi Dovesi","Matteo Poggi","Lorenzo Andraghetti","Miquel Martí","Hedvig Kjellström","Alessandro Pieropan","Stefano Mattoccia"],conference:"ICRA",year:2020,abstract:"This paper presents a compact and lightweight architecture for real-time semantic stereo matching, enabling efficient inference on embedded devices with minimal accuracy loss.",link:"https://arxiv.org/abs/1910.00541",image:"/images/papers/icra2020.png",tags:["Semantic Stereo Matching","Real-Time Processing","Deep Learning"]},{title:"Enhancing Self-Supervised Monocular Depth Estimation with Traditional Visual Odometry",authors:["Lorenzo Andraghetti","Panteleimon Myriokefalitakis","Pier Luigi Dovesi","Belén Luque","Matteo Poggi","Alessandro Pieropan","Stefano Mattoccia"],conference:"3DV",year:2019,abstract:"The authors propose a method to improve self-supervised monocular depth estimation by integrating traditional visual odometry techniques, achieving better accuracy in depth prediction.",link:"https://arxiv.org/abs/1908.03127",image:"/images/papers/3dv2019.png",tags:["Monocular Depth Estimation","Visual Odometry","Self-Supervised Learning"]}]},rm=nm.papers.filter(Qh);function lm(){const e=fr();return S.useEffect(()=>{if(e.hash==="#community"){const t=document.getElementById("community");t&&t.scrollIntoView({behavior:"smooth"})}},[e]),p.jsx("div",{className:"py-16 bg-white dark:bg-gray-800",id:"community",children:p.jsxs("div",{className:"max-w-7xl mx-auto px-4",children:[p.jsxs("div",{className:"flex items-center justify-center gap-3 mb-12",children:[p.jsx(Jc,{className:"h-8 w-8 text-green-500"}),p.jsx("h2",{className:"text-3xl font-bold text-center text-gray-900 dark:text-white",children:"Community Papers"})]}),p.jsx("div",{className:"grid md:grid-cols-2 gap-8",children:rm.map((t,n)=>p.jsxs("div",{className:"bg-gray-50 dark:bg-gray-700 rounded-xl shadow-lg overflow-hidden",children:[p.jsx("img",{className:"w-full h-48 object-cover",src:t.image,alt:t.title}),p.jsxs("div",{className:"p-6",children:[p.jsxs("p",{className:"text-sm text-green-600 dark:text-green-400 mb-2",children:[t.conference," • ",t.year]}),p.jsx("h3",{className:"text-xl font-bold text-gray-900 dark:text-white mb-2",children:t.title}),p.jsx("p",{className:"text-sm text-gray-600 dark:text-gray-300 mb-4",children:t.authors.join(", ")}),p.jsx("p",{className:"text-gray-600 dark:text-gray-300 mb-4",children:t.abstract}),p.jsxs("a",{href:t.link,className:"inline-flex items-center text-green-600 dark:text-green-400 hover:text-green-700 dark:hover:text-green-300",children:[p.jsx(Oh,{className:"h-4 w-4 mr-2"}),"View Publication"]})]})]},n))})]})})}function im(){return p.jsxs("div",{className:"pt-16",children:[p.jsx(tm,{}),p.jsx(lm,{})]})}function om(){return p.jsxs("div",{className:"pt-28 lg:pt-24 bg-white dark:bg-gray-900",children:[p.jsx("div",{className:"max-w-4xl mx-auto px-4 sm:px-6 lg:px-8",children:p.jsxs("div",{className:"text-center mb-8 lg:mb-12",children:[p.jsx($h,{className:"h-10 w-10 lg:h-12 lg:w-12 mx-auto text-green-600 dark:text-green-500 mb-4"}),p.jsx("h1",{className:"text-3xl lg:text-4xl font-bold text-gray-900 dark:text-white mb-4",children:"Manifesto"})]})}),p.jsxs("div",{className:"prose prose-base lg:prose-lg mx-auto px-4 lg:px-0 mb-16 lg:mb-20 dark:prose-invert prose-headings:text-gray-900 dark:prose-headings:text-white prose-p:text-gray-600 dark:prose-p:text-gray-300 prose-li:text-gray-600 dark:prose-li:text-gray-300 prose-h3:text-green-600 dark:prose-h3:text-green-500 max-w-4xl lg:max-w-8xl",children:[p.jsx("h3",{className:"text-xl lg:text-2xl mt-8 first:mt-0",children:"Who We Are"}),p.jsx("p",{className:"text-base lg:text-lg",children:"We are a non-profit AI research collective—scientists, professors, engineers, researchers, and developers—coming together from top universities, labs, and companies. We're here because we're curious about what AI can be and should be, and we feel a responsibility to make it count for everyone."}),p.jsx("h3",{className:"text-xl lg:text-2xl mt-8",children:"What We Believe"}),p.jsx("p",{className:"text-base lg:text-lg",children:"AI is shaping the future. It's changing every aspect of our society, including how we think about ourselves and our potential. The question is not just what AI can do, but what it should do. We believe that innovation without a guiding purpose is hollow. Real progress means more than building better tools; it means building them for everyone, especially those too often overlooked or forgotten. We want AI that doesn't just float at the top, serving the privileged, but flows downward and outward, reaching people and places long left behind."}),p.jsx("h3",{className:"text-xl lg:text-2xl mt-8",children:"What We Do"}),p.jsxs("ul",{className:"text-base lg:text-lg space-y-4",children:[p.jsxs("li",{children:[p.jsx("strong",{children:"AI Lab:"})," We focus on foundational academic AI research. We publish in open venues, share our code, and push the field forward."]}),p.jsxs("li",{children:[p.jsx("strong",{children:"Good AI:"})," We call out biases, support frameworks that keep AI honest, and build fairness, transparency, and accountability into every model we touch. We fight for meaningful, responsible applications that truly matter."]})]}),p.jsx("h3",{className:"text-xl lg:text-2xl mt-8",children:"How We Work"}),p.jsxs("ul",{className:"text-base lg:text-lg space-y-4",children:[p.jsxs("li",{children:[p.jsx("strong",{children:"Collaborations:"})," We team up with universities, research institutes, and R&D labs to tackle projects that have real impact."]}),p.jsxs("li",{children:[p.jsx("strong",{children:"Openness and Community:"})," We share what we find, learn, and create at conferences, workshops, and public events, aiming to spark honest conversation and inspire both research and action."]}),p.jsxs("li",{children:[p.jsx("strong",{children:"Partnerships:"})," Hand-in-hand with charitable non-profit organisations, we promote and develop AI applications that drive tangible improvements—education, healthcare, humanitarian aid, environmental protection, and more."]})]}),p.jsx("h3",{className:"text-xl lg:text-2xl mt-8",children:"Our Vision"}),p.jsx("p",{className:"text-base lg:text-lg",children:"We see AI as a force that reveals new paths, includes more voices, and helps reshape society for the better. If we are to contribute with a verse to this powerful play, let it be driven by curiosity and guided by responsibility."})]})]})}function am(){return p.jsx(Rh,{children:p.jsx(Nh,{children:p.jsxs("div",{className:"min-h-screen bg-white dark:bg-gray-900",children:[p.jsx(Gh,{}),p.jsxs(yh,{children:[p.jsx(Rn,{path:"/",element:p.jsx(Xh,{})}),p.jsx(Rn,{path:"/team",element:p.jsx(bh,{})}),p.jsx(Rn,{path:"/research",element:p.jsx(im,{})}),p.jsx(Rn,{path:"/manifesto",element:p.jsx(om,{})})]}),p.jsx(Yh,{})]})})})}Mc(document.getElementById("root")).render(p.jsx(S.StrictMode,{children:p.jsx(am,{})})); diff --git a/index.html b/index.html index ba81b9c..2d950c9 100644 --- a/index.html +++ b/index.html @@ -5,7 +5,7 @@ The Good AI Lab - +