This is kind of backwards…
- Continuous Image Characterization
- Continuous Image Mathematical Characterization
- Psychophysical Vision Properties
- Photometry and Colorimetry
- Digital Image Characterization
- Discrete Two-Dimensional Processing
I want to get some categories for the image processing pipeline.
- ICDAR (International Conference on Document Analysis and Recognition)
- CBDAR (International Workshop on Camera-Based Document Analysis)
- Both seem to be useless at the moment
Use the 袖珍武鑑 (shuuchinbukan.csv) books. These are 56 editions over 89 years.
I thing I need a low-level approach here since my computer is so slow. So: C++/C/Rust (don’t know how its FFI works here) Result: Did not work
- Image Improvement
- [X] Introduction
- [X] Image formation
- [X] Image processing
At least a few bullet points for each chapter of the official docs
Starting out with the OpenCV tutorials
- [X] I used ORB because not patented and from OpenCV itself. Matching looks good.
- [ ] I now like to have some metrics for comparing matching algorithms.
- [ ] Furthermore, I like to proceed with feature based alignment. Maybe building a first prototype.
- [X] Feature Detection and Matching
- [X] Segmentation
- [X] Feature-based alignment
- I this I’ll best use ORB at first
- [X] Skim rest of the book,
- [X] Especially Image-based rendering (what is this?)
- OpenCV Feature Descriptors
- I think I don’t need scale invariance; but I’ll test this!
- [X] ORB
- [X] AKAZE
- [X] BRISK
- Prepare some slides
- Ask how to best proceed
- [X] AKAZE with rotational invariance
- [X] SIFT
- [X] SURF
Just for fun. Doesn’t seem to be meaningful.
Preparing two presentations:
- [X] A general introduction of the topic
- [X] My current results
Which step produces which effect?
It’s not only about the numbers. I need to see which images succeeded and failed.
We have 366 scanned books with around 90,000 pages. Now we want to find some links for better understanding the data.
We have no ground truth!
So first, let’s apply some techniques from classical image processing.
Since this worked out better than expected, let’s see how one might use this for building a Bukan comparison platform out of this.
Seem to be solved with standard tools; just need to find the right parameters.
- Page detection
There are some current papers on this; harder than it seems but there are some working approaches.
- Page binarization
- Visualization of image changes. This shouldn’t be too sensitive to pixel changes. If thresholding and opening/closing doesn’t work I’m out of ideas. Maybe it is also possible to just paint a rectangle around some cluster of matches…
There are no (useful) existing approaches and therefore no existing tools. But it seems this isn’t the problem here. The task is too easy. ;)
- [X] Is there a difference between simple 武鑑 and 武鑑大全? Not sure, maybe just a different edition.
- [X] The Elements of Japanese Design
- [ ] A Modern History of Japan
- Do we have enough books from the same location?
- Does the number of pages match?
Do this in memory
You might want to use Histograms for finding good thresholds “Document Image Binarization” … Adaptive Thresholding
With perceptual hashes using pHash Result: Did not work!
Aligning/Registering the images and doing pixelwise comparison