Skip to main content

Posts

Showing posts with the label image-processing

OpenCV with Android

I'm just after setting up OpenCV for android in Eclipse following this tutorial , I have the samples which do not use native c running fine on my phone. The problem is when I try to run the native code samples, I followed all the steps on the next page for setting up the builder in eclipse but when I try to run the samples on my phone they crash. I have used NDK before and successfully set up tesseract for android using the command line to compile the native code instead of eclipse. I think the problem is due to unsatisfiedLinkError:Couldn't load native_sample:findLibrary returned null but I am unsure. I have a screenshot here if that helps. Any suggestions would be more than welcome !!

Render images by java code or read images from a static path?

Here's my code: 1.Java Code: public static void getImg(Long itemId) { try { Item item = Item.findById(itemId); if (item.img != null && item.img.getFile() != null{ response.setContentTypeIfNotSet(item.img.type()); renderBinary(item.img.get()); } } catch (Exception e) { Logger.error("Can't find image,itemId = " + itemId); } } html : < img src="@{{ Items.getImage(123)}}"/> 2. html : < img src="/public/images/123.jpg"/> I'm using playframework and the samples from documentation display images via the first version. What's the different (deep into mechanism) between them, especially in response performance ?

Image features extraction

I have a visual marker like this one and a blob detection algorithm in Java .. How can I extract the regions of the image so that I can run the blob detection algorithm on each one separately so that it can detect 1, 1, 3 blobs respectively. Thanks a lot in advance !

Handwritten character (English letters, kanji,etc.) analysis and correction

I would like to know how practical it would be to create a program which takes handwritten characters in some form, analyzes them, and offers corrections to the user. The inspiration for this idea is to have elementary school students in other countries or University students in America learn how to write in languages such as Japanese or Chinese where there are a lot of characters and even the slightest mistake can make a big difference. I am unsure how the program will analyze the character. My current idea is to get a single pixel width line to represent the stroke, compare how far each pixel is from the corresponding pixel in the example character loaded from a database, and output which area needs the most work. Endpoints will also be useful to know. I would also like to tell the user if their character could be interpreted as another character similar to the one they wanted to write. I imagine I will need a library of some sort to complete this project in any sort of timely

UIImage Resize without loss of quality

Im looking for an external library (or a method using Apple frameworks that doesn't destroy the quality) which will let me scale down images. I have tried this 'simple-image-processing-library' but it only works for black and white - I don't know how to make it work for colour images. ( http://code.google.com/p/simple-iphone-image-processing/ ) I have tried Core Graphics, UIKit, the "trevor" ( http://vocaro.com/trevor/blog/2009/10/12/resize-a-uiimage-the-right-way/ ) method - which makes EXTREMELY blurry images and really I don't know what else I can do - I can't continue programming my app without getting at least half-decent image scaling working. I think interpolation is needed to get decent results. Given the recent advances in iOS, I really hope someone can help me. This isn't a problem that only I am having, lots of other people seem to have the same problem, so if someone could solve it here, I'm sure if would appreciated by a