When we memorize visual stimuli, their content is processed at multiple levels, ranging from the fine-grained perceptual details to the semantic concepts and categories. However, it is unclear to which extend low- and high-level information is maintained in memory over time. Real-world stimuli are not ideal for investigating this question, as they often exhibit strong correlations between processing levels: Conceptually similar objects tend to share similar visual features. Using generative AI we created a new database of 496 image pairs orthogonalizing semantic (word2vec) and perceptual (CoreNet-S) information. Specifically, we generated image pairs that either (a) depict objects from distinct semantic concepts but are perceptually similar, or (b) show the same object but are perceptually dissimilar.