Samsung's presence at the IFA technology show in Berlin this year was suitably grandiose but most of the fun was kept for the mobile and laptop divisions, with the announcements of the Samsung Note and Samsung Galaxy Tab 7.7in, among others


According to a report coming from Taipei, Apple and its competitors are possibly implementing curved touchscreen glass displays in their line of 2012 handheld devices.


Gamespot has it from a survey done by Interpret that a significant number of Nintendo 3DS users do not like the 3D capabilities of the device.

May 02, 2015

Apple Watch Could Be a Password Alternative

With password tolerance levels at an all time low, alternatives to the pesky and insecure authenticators are beginning to abound. One of those alternatives could be the Apple Watch.

Even before Apple's latest gadget began shipping last week, MicroStrategy announced it was extending its Usher enterprise security solution to the Apple Watch.

Usher, launched earlier this year, lets an enterprise use smartphones to authenticate its users and eschew user names, passwords and tokens. The solution has two parts: a back end that runs in the cloud, and an app that runs on mobile phones.

Usher on the Apple Watch allows it to act as a digital key that employees can use to log in to business systems, unlock devices, validate an identity and open entryways.

The Usher app uses all of the Apple Watch's repertoire -- glances, notifications and force touch -- to perform its functions. It is designed to detect a variety of business systems through a wireless connection as users approach.

For example, as a user walks up to a workstation, a notification will appear on the watch to unlock the hardware, which can be done with a tap of the timepiece.

Security Plus Convenience

With Usher, an enterprise can remove the need for user names, passwords, physical IDs and badges, and replace them with a software key that can reside on a smartphone or the Apple Watch.

"That can be a transformational and important application for the enterprise, because it rids it of one of the real weaknesses in enterprise security," said Paul Zolfaghari, president of MicroStrategy.

"What the watch does is give Usher one more level of convenience," he told TechNewsWorld.

Actually, the watch may be better suited for authentication than a phone.

"The watch is designed for short interactions," Zolfaghari explained. "A short but important interaction is the transaction that takes place when somebody is being granted authorization to do something."

The watch has another benefit for the overseers of enterprise security.

"If you think about enterprise security and security in general, what you find is that for something to be more secure, it has to be less convenient," Zolfaghari said.

"It's the opposite with the Apple Watch," he continued. "You get greater security with a higher level of convenience."

Bad Password Hygiene

As if any more evidence were needed about the insecurity of passwords, LogRhythm released the latest in a long line of surveys finding that users continue to ignore good password hygiene -- except in maybe one instance.

Only 19 percent of users created unique passwords for all their accounts, according to the survey of 520 full-time employees at organizations with 250 or more employees in a number of industries.

However, when it came to their personal banking accounts, 78 percent of the respondents said they used a unique password there.

Despite the apparent concern users have over their personal finances, there seems to be little change in their general attitude toward password hygiene.

"Weak and stolen credentials is still one of the most common threat vectors used by the bad guys," said Mike Reagan, chief marketing officer at LogRhythm.

In addition to finding 81 percent of participants reused passwords in one way or another, the survey found that 79 percent of the respondents changed their passwords less frequently than once per month.

"The reality is that credentials are going to be stolen. The question is, what is the value of a credential to a bad actor?" Reagan said.

"The value of a stolen password drops precipitously when a person frequently changes passwords," he continued. "If I'm changing my password every 15 days, I'm really limiting the value of my stolen credentials."

Zombie Apps

Appthority waved a red flag last week over a what it sees as a growing problem in the mobilesphere: zombie apps.

Zombie apps are created when a program is removed from an app store but remains on a person's device. Apps can be removed for any number of reasons -- a developer pulls it because he doesn't want to support it anymore, or the operators of the store toss it for violating their terms and conditions.

"When it's removed from an app store and not from the devices, it's the equivalent of announcing a product recall and not recalling the product," said Domingo Guerra, president and cofounder of Appthority.

"It's like saying, 'We can't sell this car anymore because it's not safe,' but not telling the people who are driving the car," he told TechNewsWorld.

Because zombie apps are no longer supported by a developer, any flaws found in them by bad actors can be exploited without fear of being corrected. That can make them dangerous to a user -- even more so than mobile malware.

"We looked at over 500,000 apps on enterprise devices," Guerra said.

"Less than half a percent contained malware. More than 5 percent of iOS devices, and almost 4 percent of Android devices were dead apps," he noted.

"We also found that every single enterprise has dead apps," added Guerra. "That's not true of malware."

There would be a major corporate data breach this year that leverages a mobile device, Guerra predicted.

"We're starting to see mobile devices replacing laptops and desktops in the enterprise. The [attack] surface vector is increasing tremendously because of that," he explained. "We also seeing more and more work being done on mobile devices, so the value of the data on them is increasing."

Breach Diary

April 21. A group of small banks and credit unions request injunction against completion of US$19 million settlement announced March 19 between Target and Master Card. Group claims settlement is designed to undermine its claims against the retailer for data breach that placed at risk personal and payment card information of 103 million customers.
April 21. Senate Majority Leader Mitch McConnell introduces bill to reauthorize Section 215 of the Patriot Act through 2020 without ending blanket surveillance of Americans by U.S. intelligence agencies.
April 22. H.R. 1560, the Protecting Cyber Networks Act, passes the U.S. House of Representatives on vote of 307-116.
April 23. H.R. 1731, the National Cybersecurity Protection Advancement Act, passes U.S. House on vote of 355-63.
April 23. Adobe grants extension to the end of May to finalize a settlement of consolidated class lawsuits resulting from data breach that compromised some 3 million payment card records.
April 23. Gen. David Petraeus sentenced to two years probation and a $100,000 fine for illegally disclosing classified information to his mistress, including code-word information, detailed discussions with the president, and the names of covert operatives.
April 24. Seton Family of Hospitals in Texas informs some 39,000 patients that their personal information may be at risk following data breach discovered by the organization Feb. 26.
April 24. Congressional Budget Office estimates cost of implementing the Data Breach and Notification Act passed by Congress April 22 will be $1 million. Those costs, however, will be offset by $9 million in revenue from collection of fines imposed by the proposed law.

Courtesy; technewsworld

Follow Gadget INN' News | Reviews on Twitter:

October 22, 2011

Canon EF 100mm f/2.8 USM Macro Review

The Canon EF 100mm f/2.8 USM Macro is one of two new 100mm macro lenses in Canon's range, the other being the more expensive L-series optic, the CanonEF 100mm f/2.8L Macro IS USM, which has the advantage of a new Hybrid IS system.

This one may lack image stabilisation but, nevertheless, it's priced attractively and competitively specified.

It contains an internal focusing system, which has the dual advantages of maintaining the same working distance throughout its focusing range and keeping the front element static as this happens.

Focusing itself is carried out via a ring Ultrasonic Motor (USM), which Canon claims offers fast and quiet autofocus, while the further inclusion of a three-group floating system is said to ensure correction for aberrations throughout the lens's focusing range.

On the inside, Canon has incorporated a secondary diaphragm to block stray light when shooting at f/2.8, while a single Ultra Low Dispersion element is also included to help minimise chromatic aberrations.

To help improve light transmission and to reduce both flare and ghosting, Canon has treated the elements within the lens with its Super Spectra Coatings. This works by cancelling out the reflections from the surfaces of the elements, so that more light can pass through to the sensor and not bounce around inside to degrade image quality.

The lens incorporates a particularly broad rubber manual focusing ring on its barrel, which turns easily and makes clear when the ends of its focusing scale have been reached. It can also be used to override the autofocus system once focus has been confirmed.

This is joined by two switches next to the distance window - one that alternates between manual and autofocus and another that closes off the focusing range to 0.48m-infinity for non-macro photography.

The distance window, meanwhile, indicates the focusing distance in both feet and metres, and also shows the current reproduction ratio down the maximum 1:1.

The mottled finish of the Canon EF 100mm f/2.8 USM Macro's body is similar to that of some previous Canon EOS DSLR cameras, such as the EOS 600D and EOS 5D Mk II, although its design is a little dated when compared with more recent finish options from other manufacturers.

The mount is constructed from a solid metal for durability, and above it sits a large red lens mount index that indicates the mounting position clearly.

The only other external features of note are the 58mm filter thread at the opposite end of the lens and the grooves around it, which can be used to mount either an external lens hood or a macro ring lite flash.

At this price we don't necessarily expect an image stabilisation system, and for macro work a sturdy tripod is generally recommended as a better idea. However, since Canon lenses don't offer image stabilisation inside their bodies, this puts it at a slight disadvantage to similar lens/body combinations from other manufacturers when used for non-macro work.

This is particularly the case with APS-C bodies such as the Canon EOS 600D and EOS 7D, where its effective focal length rises closer to 160mm, and this omission also means that the viewfinder image isn't stabilised for more precise composition.


Thanks to the use of a piezoelectric motor, focusing noise is about as silent as expected as the lens travels between its two extremes, and it's able to quickly shift its focus between varying distances before promptly confirming focus.

Furthermore, the use of different text colours in the distance window makes it easy to differentiate between its three readings, while the size of the focusing ring means that it's easily adjusted with just a finger if required.

Its positioning towards the front of the barrel makes this manoeuvre a little more difficult for those with smaller hands, but in most situations the macro lens will no doubt be either left to autofocus or be manually adjusted more precisely on a tripod.

Control over distortion is good, and the absence of chromatic aberration in the centre of the frame is particularly impressive, although as you venture further away from the centre it's clear that there's room for improvement.

Although central sharpness at f/2.8 is impressive this slowly drops off to softness towards the edges. By f/8, however, all is forgiven as the Canon EF 100mm f/2.8 USM Macro records excellent sharpness right across the frame. Predictably, things aren't quite as good at its smallest apertures, although even here consistency is still reasonable.


Overall then, this lens is a capable performer, with quick and silent autofocus system and generally good to excellent image quality. While more expensive then the Tamron SP AF 90mm F/2.8 Di Macro 1:1, it has the distinct advantages of an internal focusing system and a much quieter motor.

(Reviewed by Matt Golowczynski - TechRadar )

Follow Gadget INN' News | Reviews on Twitter:

HDR :Better results from multiple exposures

The sensor in a modern digital camera has less dynamic range than the human eye. That's why we're often disappointed with photographs we take: we don't see the sky as washed out, or the shadows as dark as they appear in our photos.

Naturally, there are now ways to circumvent this using the power of the PC: enter the high dynamic range image processing algorithms.

When we purchase a digital camera, we're often concerned with the resolution of the sensor (the number of megapixels), whether it produces images in JPG or RAW format, and whether we can use different lenses to get images from close up or far away. We're not generally concerned with the dynamic range of the sensor in the camera – in other words, the range of light levels that the sensor can capture.

It turns out that old-style film has less dynamic range than a CCD (charge-coupled device) – the sensor that registers light information that's built into modern digital cameras. If you like, we've moved forward in terms of dynamic range and also, incidentally, in terms of noise: film is noisier at low light intensities than digital.

EYE VS SENSOR: How the eye perceives light intensity (blue line) compared to the way the camera does (red line)

Both types of camera perceive a narrower range of light levels than the human eye. That's why when we take a photograph of a landscape, for example, we get an image that doesn't register the cloud formations in the sky (the sky becomes washed out) and the shadows become undifferentiated black.

If you're anything like me, you tend to get a little frustrated and disappointed that the camera isn't recording what you're seeing exactly, but I dare say we've all become rather used to the problem.

What is HDR?
There is, however, a way around this called HDR image processing (high dynamic range). This is a set of algorithms that process images to increase their dynamic range.

With HDR, it's possible to produce an image that has a much greater range of light levels to approximate what the human eye can see, or even to make fantasy images that look nothing like real life.

However, I'll also note that you will run into some issues. For example, the monitors we use to view images also have a smaller dynamic range than the human eye.

Back in the days of film cameras, it was possible to increase the dynamic range of a photo when you printed the image after developing the film. Photographers like Ansel Adams were experts at using this kind of image manipulation – known as dodging and burning – to produce the dramatic photos we've all seen and perhaps bought as posters.

Dodging decreases the exposure of the print making the area lighter in tone, whereas burning increases it making the area darker in tone. Recall that in film photography, the film is a negative version of the photo. Dark areas on the negative will show up as lighter areas on the print paper because the light-sensitive silver salts in the paper will be less exposed, and therefore appear lighter once the print is fixed. Light areas will show up darker on the print, because more light hits the silver salts.

To apply dodging to the print, the photographer would cut out a shape from some opaque material like card to block off part of the scene, and then expose the print paper with that card between the projector and the paper. Because less light hit that part of the scene, it would appear lighter.

Burning was done in a similar way, but this time the photographer wanted to expose part of the scene longer than the rest. They would cut out a shape that would block off the rest of the scene, letting the part to be brightened receive more light.

There are other techniques and materials that can be used, but as you can imagine, dodging and burning this way was a labour-intensive process and was usually only done for art photos and the like.

Another issue is that dodging and burning are physical manipulations that happen in real time, and it's hard to replicate the same effects across a set of prints so that the resulting images are all the same. With digital photographs and programmatic image manipulation, it's a lot easier to create images with a higher dynamic range.

The process goes like this. First you take at least three photographs of a scene. Ideally, these photos are taken with your camera on a tripod so that they all register exactly the same scene. I've tried using a hand-held camera and it just doesn't work as well during the HDR processing – there's always some obvious scene shake that can be seen.

Similarly, the scene itself should be as static as possible: any moving parts (like leaves fluttering in the wind, waves crashing on the beach, or cars or people passing by, for example) won't be the same in each image, causing scene shake in the processed HDR file.

Although the photos are of the same scene and have the same focus and aperture settings, you take them using different exposure times. For best results, you should shoot one photo as normal, and the others two stops either side. Increasing exposure by a single stop doubles your exposure time, and decreasing by a stop halves it.

The dynamic range in photography is the number of stops between the darkest part of an image where you can still resolve detail and the lightest part. DSLRs generally have about 11 stops of dynamic range at low ISO values, and point-and-shoot cameras a stop or so less.

The external Viewsonic monitor I'm using with my MacBook Pro has almost 10 stops, which means that the photos I take with my camera already have twice the dynamic range that my screen can show.

Some cameras, especially DSLRs, have a mode whereby you can shoot three photos as a set, the other two bracketing the first in terms of exposure. The different exposures are regulated by the camera automatically.

On my Canon Rebel XTi (also known as the 400D), this is AEB mode (auto-exposure bracketing) and I can set the required +/- 2 stops there. If your camera shoots RAW instead of JPG and your HDR image processing application supports it, it's possible to just use one photo. The results won't be nearly as good though.

Once you have your three differently exposed photos, it's time to process them. The first stage is to analyse all three photographs in order to merge them as a single HDR image.

Encoding photos

NORMAL EXPOSURE: Our scene normally exposed

Without getting bogged down in compression details, image formats and so on, traditional images (like an individual photo from our set of three) encodes colour information for each pixel as a set of three bytes, one for red, one for green and one for blue. Each colour channel for each pixel can therefore represent 256 different levels, and the pixel itself can be defined as a single 24-bit value.

An HDR image is different. In a simple sense, it encodes more bits per colour channel per pixel than a standard image, but that's not the entire story. To understand why, we need to understand what's meant by 'gamma correction'.

OVER-EXPOSED: Our scene over exposed. Notice the difference in colour tones to the normal exposure

Gamma defines the difference between a pixel's colour values and how bright it is actually perceived (its luminance). For a camera, if you double the amount of light on a sensor's pixel, a value twice the original is detected – the relationship between a pixel's value and its luminance is linear.

This linear relationship doesn't apply for our eyes. When we increase luminance at low light levels, we perceive a larger increase in light. At higher light levels, we don't perceive increases in luminance as well. Our eyes are more sensitive to changes in dark tones than changes in light tones.

Accounting for this difference is known as gamma correction. A camera will apply gamma correction to an image before it saves it as a JPG file. It's the same with image processing applications that work with RAW image files, when you save an image as a JPG. In other words, instead of encoding the values of the pixels reported by the sensor, it will apply a gamma correction (the industry standard value is 1/2.2) first.

Instead of recording the real values of the pixel colours, it will encode them as colour tones that we perceive as varying uniformly. In effect, the code that saves an image as a JPG file uses more bits to encode darker colours than lighter ones.

When you view such a gamma-encoded JPG on your screen, the screen software will apply the reverse gamma correction (a gamma of 2.2 usually) so that the image you see is roughly the same as the original scene. Lighter shades will have less variation than darker shades, though.

UNDER-EXPOSED: Again, notice the differences in colour between this and the first two shots

Notice also that merging three JPGs to produce an HDR image would necessitate that the gamma correction be reversed before the merge. JPG is already a lossy compression algorithm (we're losing image information when the camera creates a JPG, smearing out high-variation regions of the image), which means that our three differently-exposed photos should all be in RAW format.

Now back to the HDR image. It converts colours as tones using the gamma correction just discussed (so darker tones have more shades) and encodes them as a floating point number, using either single precision (32-bit) or double precision (64-bit). All three standard images are used in this conversion/encoding process, since each will have a different complementary set of tone information.

There's a lot of redundant information, so the most popular HDR format compresses the RGB tone data as three fixed point values, with an extra byte holding a common exponent.

Viewing the image

TONE-MAPPED HDR: And like magic, the final result shows more dynamic colours and contrast to the original photos

So now we have an HDR image whose tones are encoded in high fidelity using a special encoding scheme. To view it, however, we have to export it as a JPG again.

Because of the +/- 2 stops bracketing, the HDR image has a dynamic range of roughly 14 stops, which happens to be about as much as the eye can perceive, but we have to convert that wide dynamic range into the 10 stops of the screens we use.

We have to map the HDR tones into the range that can be displayed on a monitor. Since we're going to be removing information (or reducing the amount of data) in order to do this, tone mapping is a lossy one-way process. It also tends to be an interpretive process, much as dodging and burning was to the film photographers, requiring a good eye and attention to detail and lots of time tweaking and experimenting.

In any HDR image-processing application, there will be several knobs or sliders you can adjust to change the tone mapping, and thereby the look and feel of the finished low dynamic range image (sometimes known as LDR).

Some examples include exposure or brightness, gamma or contrast – sharpening tools that alter the local appearance of parts of the image. Some may come with presets that generate an LDR image according to the image style you want to show (Photomatix, the application I use, has a preset called Grunge and yes, the results are pretty grungy).

There are several tone-mapping algorithms that can process an HDR image to produce an LDR version (a hands-off algorithm), but it's a field that is being actively researched and is changing rapidly.

The biggest issue that automated algorithms have to solve is that tone mapping, by its very nature, reduces contrast (contrast is a difference in luminance: the higher the contrast, the greater the difference in luminance or tone). The earliest hands-off algorithms were global algorithms that acted on the image as a whole.

An example of global tone mapping is to take the tone (or luminance) information and scale it linearly to the required range. Luminance is a value on an exponential scale (each increase of one stop doubles the brightness), so the linear scaling is usually done on the logarithm of the tone values. The problem with this was that it paid no attention to areas within the image with local high contrast.

Other algorithms attempt to take into consideration regions of the image that have wide variations in contrast. The algorithms use a range of methods to generate an LDR image as a multi-pass process. Some, for example, tone map most of the image one way, then map the localised high-variation regions using another method to maintain contrast.

As you can see, the problem of HDR image processing hasn't been completely solved yet. Nevertheless, it's fast becoming an accepted way to process images with high contrast, and can produce some stunning photographs.

Follow Gadget INN' News | Reviews on Twitter:

October 20, 2011

BlackBerry App News: Spotify finally arrives on App World

After yesterday's apologetic bounty of free apps, RIM has some good news on the app front today, with the announcement that Spotify is finally available for the BlackBerry platform. Finally.

Spotify has finally unleashed its Premium music streaming services on the BlackBerry-toting masses, by launching its mobile app.

Spotify mobile is already available for iOS, Android, Windows Phone, Nokia and even WebOS got some love from the Swedes before the BlackBerry App World is now getting a look in.

To make use of the app, just like on all of those aforementioned platforms, you'll need a Spotify Premium subscription, which costs a tenner a month.

Once you're signed up you'll have access to millions of songs, offline playlists stored on your handset and syncing of playlists between desktop and mobile versions. There's also track scrobbling to

The app is available to download in preview from the BlackBerry App World, but a full and complete version will be released very soon, according to RIM.

The launch of Spotify comes at an ideal time for RIM, which is looking to convince current users that it is worth staying loyal to the brand following last week's disastrous internet services outage which brought BBM down across the world.

Earlier in the week, the company says it was offering £63 worth of premium apps free for its users as an apology for the whole sorry episode.

Link: TechRadar

Follow Gadget INN' News | Reviews on Twitter:

Related Posts Plugin for WordPress, Blogger...