Growing concerns about data privacy and the legality of data sourcing by major AI companies such as OpenAI and Google have intensified in recent years. Against this backdrop, Sony’s newly released Fair Human-Centric Image Benchmark (FHIBE, pronounced “Fee-bee”) offers a refreshing and responsible alternative.
FHIBE is the first publicly available dataset featuring humans that was created under strict ethical standards — including informed consent, participant compensation, cultural diversity, and full transparency in the annotation process. Initiatives like this are essential for ensuring that AI models are trained and tested in a safe, lawful, and transparent way.
The dataset contains over 10,000 images from nearly 2,000 individuals across 81 countries and regions. It supports tasks such as facial recognition, pose estimation, silhouette segmentation, and visual question answering (VQA). Crucially, all images were captured specifically for this project, with participants giving explicit consent for their use in AI research. Unlike many existing datasets, FHIBE does not rely on scraped or archival internet material, thereby reducing privacy risks and enabling more accurate benchmarking of AI models.
Researchers note that many current computer vision systems are trained on data lacking ethnic, age, and cultural diversity, which leads to biased or inaccurate results in medical, industrial, and public safety applications — especially for underrepresented groups. FHIBE aims to close this gap by providing scientists and technology companies with a tool to objectively test models for bias and ethical compliance.
The authors emphasize that FHIBE not only improves algorithmic accuracy but also sets a new standard for data science. Responsible data collection, transparency, and public verifiability are becoming as important as neural network architectures themselves.
The AI industry is already facing mounting legal and ethical challenges, and we are still at the beginning of its evolution. It’s hard to imagine the consequences if AI models were to continue being trained for the next 5–10 years without oversight or regulation. That’s why FHIBE represents such an important step — it anticipates upcoming regulations and establishes a new, ethically sound benchmark for developing and training AI models responsibly.

