Search Results

Search found 24784 results on 992 pages for 'process integration packs'.

Page 462/992 | < Previous Page | 458 459 460 461 462 463 464 465 466 467 468 469  | Next Page >

  • Multiple volumetric lights

    - by notabene
    I recently read this GPU GEMS 3 article Volumetric Light Scattering as a Post-Process. I like the idea to add volumetric light property to realtime render i'm working on. Question is will it work for multiple lights? Our renderer uses one render pass per light and uses additive blending to sum incoming light. I'm mostly convinced that it have to work nice. Do you agree? Maybe there can be problem where light rays crosses each other.

    Read the article

  • Clonezilla failing as soon as image copying begins

    - by mmr
    I have been trying unsuccessfully to create an image of an Ubuntu 10.04 laptop system. As soon as the copying itself starts, the entire system crashes to a black screen. I suspect that the problem is overheating, and that's why I put an ice pack under the machine. That seems to have helped a bit, but it's still not getting through the copying process. Is there any other possible explanation for dying to a black screen like this? I'm just not relishing the task of removing the hard drive, mounting it elsewhere, and then doing a backup that way.

    Read the article

  • Did I back that file up ?

    - by GrumpyOldDBA
    As part of some diagnostics I was wishing to check if a particular file had been locked by a backup process whilst I was trying to copy it. So I put a request into the DataCentre for a list of files backed up and time when they were backed up, seems reasonable to me ? However the response was somewhat strange, apparently I can only find out if a file was backed up if I ask to have it restored, there is no facility ( so I'm told ) to produce a list of backed up files. It's good to know that...(read more)

    Read the article

  • What is a good use case for scala?

    - by Usman Ismail
    In a current project we have setup the build so that we could mix Java and Scala. I would like to use more Scala in our code base to make the code more readable and concise. In the process also learn the language by handing over real features. So I plan to use Scala for some classes to showcase its benefits and convince other devs to look into using Scala too. For a rest based web server or a program in general what kind of code structures lend themselves to Scala's functional programming style.

    Read the article

  • Point domain to 3rd Party DNS

    - by PhilCK
    I have a few of domain names and a rather simple website (small company type thing). We are in the process of having a web designer create a new website for us, but I don't want to give access to the control panel for the domain names (and have no way to limit it, it seems), while at the same time I don't want to be the go between guy for it the settings. Is there a way or a service for me to point the domain's at a 3rd party DNS system, that I can then give access for the web designer, without worry that he can find my personal info or try and transfer my domain out. Thanks.

    Read the article

  • Installation of Ubuntu 13.04 seems to take forever

    - by Michel
    I am about to install Ubuntu 13.04 on my AMD x64 machine and the installation of the lib packages seems to take forever. Is that a known issue? I know that it might take some time when checking to download packages from the Internet but this went through very fast. In the terminal-like view I can see, the system is just unpacking, installing, configuring etc. in a quite slow way. In the opening screen it reads: It just takes a few minutes (or something close to that). Again: Is that a normal behavior or am I doing something wrong and what could I do to speed up the process in case I'm running an installation again on another system?

    Read the article

  • How to create an auto-grader in and for Python

    - by recluze
    I'm trying to create an auto-grader for one of my beginning programming courses for python. From my online search, I've come to know that it is effectively a unit test framework that tests the student's code rather than production code but I'm not really sure how to structure the flow of the program. Can anyone please provide a strategy for submission of code by students and automating the whole process of marking? For instance, how would the student code be submitted and then stored/structured on disk, how would the grades be stored/reported? I'm only looking for a broad strategy and will try on my own to fill in the blanks. (I asked this on stockoverflow.com initially but it's considered as off-topic and I was suggested to ask here.)

    Read the article

  • IKEA Lamps Hacked into Flexible Speaker Mounts

    - by Jason Fitzpatrick
    This simple hack combines the swing arms of two IKEA work lamps with a set of computer speakers for flexible and easily adjustable sound. IKEAHackers reader Bill Dwyer wanted an easy way to get the speakers off his desk but still be able to easily adjust them. By hacking apart two IKEA work lamps (he removed the light assembly and snipped the wires off) he was able to attach his computer speakers to the arms and, in the process, get them off the desk. The arms make it super simple to adjust the speakers exactly where he wants them, including towards other parts of his office/apartment. Hit up the link below to check out more pictures and read Bill’s instructions. Very Flexible Computer Speaker Mounts [IKEAHackers] Use Your Android Phone to Comparison Shop: 4 Scanner Apps Reviewed How to Run Android Apps on Your Desktop the Easy Way HTG Explains: Do You Really Need to Defrag Your PC?

    Read the article

  • Disaster Recovery Discovery

    - by Rodney Landrum
    Last weekend I joined several of my IT staff on a mission to perform a DR test in our remote CoLo center in a large South East city of the US. Can I be more obtuse? The goal was simple for me as the sole DBA in a throng of Windows, Storage, Network and SAN admins – restore the databases and make them work. There were 4 applications that back ended to 7 SQL Server databases on 4 different SQL Server instances. We would maintain the original server names, but beyond that it was fair game. We had time to prepare so I was able to script out or otherwise automate the recovery process. I used sp_help_revlogin for three of the servers, a bit of a cheat actually because restoring the Master database on the target DR servers was the specified course of action according to the DR procedures ( the caveat “IF REQUIRED” left it open to interpretation. I really wanted to avoid the step of restoring Master for a number of reasons but mainly because I did not want to deal with issues starting SQL Services afterward. Having to account for the location of TempDB and the version conflicts of the resource DBs were just two of the battles I chose not to fight. Not to mention other system database location problems that might arise and prevent SQL from starting.  I was going to have to restore all of the user databases anyway, so I would not really gain any benefit, outside of logins, for taking the time to restore the source Master database over the newly installed one on the fresh server. What I wanted was the ability to restore the Master database as a user database, call it Master_Mine, from a backup on the source system and then use that restored database to script the SQL Logins and passwords on the DR systems. While I did not attempt this on the trip, the thought stuck in my mind and this past week I succeeded at scripting user accounts and passwords using only a restored copy of the Master database. Granted there were several challenges to overcome.  Also, as is usual for any work like this the usual disclaimers apply:  This is not something that I would imagine Microsoft would condone or support and this was really only an experiment for me to learn if it was even possible. While I have tested the process with success, I do not know that I would use this technique in a documented procedure because future updates for SQL Server will render this technique non-functional. I thought at first, incorrectly of course, that I could use sp_help_revlogin on a restored copy of the master database I named Master_Mine.   Since sp_help_revlogin uses system schema objects, sys.syslogins and sys.server_principals, this was not going to work because all results would come from the main Master database. To test this I added a SQL login via SSMS, backed up Master, restored  it as Master_Mine, and then deleted the login.  Even though the test account I created should presumably still be in the Master_Mine database, I should be able to get to it and script out its creation with its password hash so that I would not need to know the password, but any applications that stored that password would not have to be altered in the DR scenario. They would just work as expected. Once I realized that would not work I began looking deeper.  Knowing that sys.syslogins and sys.server_principals are system views, their underlying code should be available with sp_helptext, right? They were. And this led me to discover the two tables sys.sysxlgns and sys.sysprivs, where the data I needed was stored. These tables existed in both the real Master and the restored copy, Master_Mine.  I used this information to tweak the sp_help_revlogin stored procedure to use these tables instead to create the logins cursor used in sp_help_revlogin. For the password hash,  sp_help_revlogin uses the function LoginProperty() which takes a user name and option ‘passwordhash’ to return the hash for the user. Unfortunately, it requires the login to exist in the Master database. This would not work. So another slight modification I had to make was to pull the password hash itself (pwdhash from sys.sysxlgns) into the logins cursor and comment out the section of sp_help_revlogin that uses LoginProperty. Instead, I pass the pwdhash value as the variable @PWD_varbinary to the sp_hexadecimal stored procedure which is also created by and used within the code provided by Microsoft in the link above for sp_help_revlogin. The final challenge: sys.sysxlgns and sys.server_principals are visible only within a Dedicated Administrator Connection (DAC) query window in SSMS or within SQLCDMD.  To open a DAC connection you have to be logged in on the SQL Server itself, via RDP in my case,  and you preface the server name in the query connection with ADMIN:, so that the server connection looks like ADMIN:ServerName. From there you can create the modified stored procedure in the restored copy of a Master database from a source system as whatever name you like, and then run the modified stored procedure. I named my new stored procedure usp_help_revlogin_MyMaster. Upon execution I was happy to see the logins and password hashes that I needed to apply from the source Master database without having to restore over the new Master system database and without the need to access the original server (assuming it was down due to whatever disaster put it in that state). You will note that I am not providing full code samples here of the modifications. I will say that it was a slight bit of work and anyone who needed to do this for whatever reason, could fairly easily roll their own solution with the information provided herein.  My goal, as I said was to prove that this could be done and provide another option if required to ease the burden of getting SQL Servers up and available in an emergency situation where alternatives may be more challenging or otherwise unavailable.  

    Read the article

  • Can't boot freshly burned Ubuntu cd

    - by user89004
    So I just burned a Ubuntu 12.04.1 powerpc .iso on a cd for my iMac G5 running Mac OS X 10.4.11 and it won't even recognize the cd. I burned it on my dad's Windows 7 laptop as the process is way easier (just 2 clicks). Mac OS X 10.4.11 gives me an error when it starts and when the CD is in saying "the disk you inserted was not readable by this computer". What's funny is that I burned a Ubuntu Minimal .iso on a CD and it would totally read that and even boot it though it gave me some errors afterwards and I couldn't install. I even tried going into openfirmware and hitting boot cd:,\tbxi but I get the error "Warning sector size mismatch can't OPEN cd:,\tbxi Can't open device or file" Was there something wrong with the .iso I burned? Mac OS X 10.4.11 won't even mount that .iso it tells me that the HFS file system is corrupt or something, but I know the .iso doesn't contain HFS file system. Any help? I downloaded the .iso from http://cdimage.ubuntu.com/releases/precise/release/ubuntu-12.04-desktop-powerpc.iso

    Read the article

  • Ubuntu 12.04.1 LTS breaking down and sluggish

    - by ahamza
    I am in the process of switching from Windows to a Linux based system and am currently deciding between distributions...I am currently trying Ubuntu as WUBI. I find that my experience is not very smooth and streamlined for example, crash-reports, bugs, applications taking time to run etc etc (this in spite of the fact that I am very patient and am constantly researching solution to different driver and application issues). Was wondering if this is because I am running through NTFS right now or is it just like this? Looking to switch to Linux because of its opensource nature, interest in software development in college as well as maximizing the potential of my machine. I am running an AMD quadcore-x64 2.2GHz, 6GB RAM and 750 GB HDD on an HP G6 notebook. I would appreciate any honest opinions.

    Read the article

  • Big Data – Buzz Words: What is HDFS – Day 8 of 21

    - by Pinal Dave
    In yesterday’s blog post we learned what is MapReduce. In this article we will take a quick look at one of the four most important buzz words which goes around Big Data – HDFS. What is HDFS ? HDFS stands for Hadoop Distributed File System and it is a primary storage system used by Hadoop. It provides high performance access to data across Hadoop clusters. It is usually deployed on low-cost commodity hardware. In commodity hardware deployment server failures are very common. Due to the same reason HDFS is built to have high fault tolerance. The data transfer rate between compute nodes in HDFS is very high, which leads to reduced risk of failure. HDFS creates smaller pieces of the big data and distributes it on different nodes. It also copies each smaller piece to multiple times on different nodes. Hence when any node with the data crashes the system is automatically able to use the data from a different node and continue the process. This is the key feature of the HDFS system. Architecture of HDFS The architecture of the HDFS is master/slave architecture. An HDFS cluster always consists of single NameNode. This single NameNode is a master server and it manages the file system as well regulates access to various files. In additional to NameNode there are multiple DataNodes. There is always one DataNode for each data server. In HDFS a big file is split into one or more blocks and those blocks are stored in a set of DataNodes. The primary task of the NameNode is to open, close or rename files and directory and regulate access to the file system, whereas the primary task of the DataNode is read and write to the file systems. DataNode is also responsible for the creation, deletion or replication of the data based on the instruction from NameNode. In reality, NameNode and DataNode are software designed to run on commodity machine build in Java language. Visual Representation of HDFS Architecture Let us understand how HDFS works with the help of the diagram. Client APP or HDFS Client connects to NameSpace as well as DataNode. Client App access to the DataNode is regulated by NameSpace Node. NameSpace Node allows Client App to connect to the DataNode based by allowing the connection to the DataNode directly. A big data file is divided into multiple data blocks (let us assume that those data chunks are A,B,C and D. Client App will later on write data blocks directly to the DataNode. Client App does not have to directly write to all the node. It just has to write to any one of the node and NameNode will decide on which other DataNode it will have to replicate the data. In our example Client App directly writes to DataNode 1 and detained 3. However, data chunks are automatically replicated to other nodes. All the information like in which DataNode which data block is placed is written back to NameNode. High Availability During Disaster Now as multiple DataNode have same data blocks in the case of any DataNode which faces the disaster, the entire process will continue as other DataNode will assume the role to serve the specific data block which was on the failed node. This system provides very high tolerance to disaster and provides high availability. If you notice there is only single NameNode in our architecture. If that node fails our entire Hadoop Application will stop performing as it is a single node where we store all the metadata. As this node is very critical, it is usually replicated on another clustered as well as on another data rack. Though, that replicated node is not operational in architecture, it has all the necessary data to perform the task of the NameNode in the case of the NameNode fails. The entire Hadoop architecture is built to function smoothly even there are node failures or hardware malfunction. It is built on the simple concept that data is so big it is impossible to have come up with a single piece of the hardware which can manage it properly. We need lots of commodity (cheap) hardware to manage our big data and hardware failure is part of the commodity servers. To reduce the impact of hardware failure Hadoop architecture is built to overcome the limitation of the non-functioning hardware. Tomorrow In tomorrow’s blog post we will discuss the importance of the relational database in Big Data. Reference: Pinal Dave (http://blog.sqlauthority.com) Filed under: Big Data, PostADay, SQL, SQL Authority, SQL Query, SQL Server, SQL Tips and Tricks, T SQL

    Read the article

  • Do logged in users need to browse a site over https?

    - by Luke
    I've never thought it was necessary, but a client has requested that all webpages served to logged in users be delivered over HTTPS. Aside from the implementation standpoint, which I don't think I'm going to pursue is there any real reason for this request ? For clarity, the login / logout process, account settings, registration preferences and all user related scripts are served over https. but I can't see the point in my news articles, press releases, events etc... being served in this manner? Am I missing something ?

    Read the article

  • Are There Realistic/Useful Solutions for Source Control for Ladder Logic Programs

    - by Steven A. Lowe
    Version control for ladder logic (LL) programs for programmable logic controllers (PLCs) seems to be virtually non-existent. It may be because LL is a visual language and tends to be stored in binary files, or it may be because source code control hasn't "caught on" in process control engineering circles - or perhaps my Google-Fu is weak tonight. Do you know of any realistic and useful solutions for version control for such systems? Definitions: realistic = changes to the programs are tracked by user and subject to reversion and merges useful = the system integrates with visual LL designers, is not limited to LL from a single PLC manufacturer, and does not cost a ridiculous amount of money? Note: I have heard of people using SVN or Mercurial et al to track the binary files, but I don't think the diff/merge capabilities would display readable differences.

    Read the article

  • Compelling Reasons For Migrating to Oracle Database 11g

    - by margaret hamburger
    IDC's white paper Maximizing Your Investment in Oracle Application Software: The Case for Migrating to Oracle Database 11g  describes compelling business, operational, and technical reasons for Oracle application customers to upgrade or Oracle Database 11g. In researching this paper, IDC found that the upgrade process is smooth, the latest version offers key benefits over older versions, and the new features and procedures are easy to learn and well worth implementing. The paper highlights users of Oracle applications, such as Oracle's PeopleSoft Enterprise applications, Siebel Customer Relationship Management (CRM) applications, and Oracle E-Business Suite. It includes a review of Oracle Database 11g improvements and new features along with best practices from Oracle users who upgraded to Oracle Database 11g.

    Read the article

  • How to customise window decoration whilst using Compiz on Xubuntu?

    - by Benjamin
    I have installed Compiz on Xubuntu 11.10 with sudo apt-get install compiz compizconfig-settings-manager compiz --replace ccp & In the process the XFCE window decoration theme is overridden by that of Compiz (Gtk) which uses the Adwaita theme instead of the Greybird theme. Since Gtk is doing window decoration, I cannot change it back using the XFCE settings. I just need compiz for scale and window switch and I would like to return window decoration to XFCE (Xfwm4) or to be able to change the Gtk window decoration theme. How can I do that? I have found part of the (workaround) answer already: download Greybird Gtk theme install theme (here is where I failed I think) use dconf-editor to change the Gtk theme in org.gnome.desktop.interface The problem really at stage 2 is where do I place the theme? I tried in ~/.themes/ and then changed the value of gtk-theme in the editor to Greybird. But I saw no change.

    Read the article

  • 3D Display Issue When Using Latest Java Runtime Versions - Patch now available...

    - by [email protected]
    Typically I focus my blog posts on Support process topics, and reserve most of the technical topics for the Support newsletter. This topic, however, warrants a quick mention in the blog since I know it's been affecting many users recently. For customers using the Client/Server Deployment of AutoVue, users that had upgraded their client Java Runtime Environment (JRE) to version 1.6.0_19 or later suddenly noticed that their 3D files were opening blank in AutoVue. This issue was due to a change in JRE version 1.6.0_19, and the AutoVue team now offers a patch to address the issue in AutoVue version 20.0.0. The patch number is 10268316, is available through the My Oracle Support portal, and is described further in KM Note 1104821.1. We'll mention it again in our next Support newsletter, and the AutoVue team will target to roll the same fix into the next available release of the product.

    Read the article

  • JavaOne 2012 Sunday Strategy Keynote

    - by Janice J. Heiss
    At the Sunday Strategy Keynote, held at the Masonic Auditorium, Hasan Rizvi, EVP, Middleware and Java Development, stated that the theme for this year's JavaOne is: “Make the future Java”-- meaning that Java continues in its role as the most popular, complete, productive, secure, and innovative development platform. But it also means, he qualified, the process by which we make the future Java -- an open, transparent, collaborative, and community-driven evolution. "Many of you have bet your businesses and your careers on Java, and we have bet our business on Java," he said.Rizvi detailed the three factors they consider critical to the success of Java--technology innovation, community participation, and Oracle's leadership/stewardship. He offered a scorecard in these three realms over the past year--with OS X and Linux ARM support on Java SE, open sourcing of JavaFX by the end of the year, the release of Java Embedded Suite 7.0 middleware platform, and multiple releases on the Java EE side. The JCP process continues, with new JSR activity, and JUGs show a 25% increase in participation since last year. Oracle, meanwhile, continues its commitment to both technology and community development/outreach--with four regional JavaOne conferences last year in various part of the world, as well as the release of Java Magazine, with over 120,000 current subscribers. Georges Saab, VP Development, Java SE, next reviewed features of Java SE 7--the first major revision to the platform under Oracle's stewardship, which has included near-monthly update releases offering hundreds of fixes, performance enhancements, and new features. Saab indicated that developers, ISVs, and hosting providers have all been rapid adopters of the platform. He also noted that Oracle's entire Fusion middleware stack is supported on SE 7. The supported platforms for SE 7 has also increased--from Windows, Linux, and Solaris, to OS X, Linux ARM, and the emerging ARM micro-server market. "In the last year, we've added as many new platforms for Java, as were added in the previous decade," said Saab.Saab also explored the upcoming JDK 8 release--including Project Lambda, Project Nashorn (a modern implementation of JavaScript running on the JVM), and others. He noted that Nashorn functionality had already been used internally in NetBeans 7.3, and announced that they were planning to contribute the implementation to OpenJDK. Nandini Ramani, VP Development, Java Client, ME and Card, discussed the latest news pertaining to JavaFX 2.0--releases on Windows, OS X, and Linux, release of the FX Scene Builder tool, the JavaFX WebView component in NetBeans 7.3, and an OpenJFX project in OpenJDK. Nandini announced, as of Sunday, the availability for download of JavaFX on Linux ARM (developer preview), as well as Scene Builder on Linux. She noted that for next year's JDK 8 release, JavaFX will offer 3D, as well as third-party component integration. Avinder Brar, Senior Software Engineer, Navis, and Dierk König, Canoo Fellow, next took the stage and demonstrated all that JavaFX offers, with a feature-rich, animation-rich, real-time cargo management application that employs Canoo's just open-sourced Dolphin technology.Saab also explored Java SE 9 and beyond--Jigsaw modularity, Penrose Project for interoperability with OSGi, improved multi-tenancy for Java in the cloud, and Project Sumatra. Phil Rogers, HSA Foundation President and AMD Corporate Fellow, explored heterogeneous computing platforms that combine the CPU and the parallel processor of the GPU into a single piece of silicon and shared memory—a hardware technology driven by such advanced functionalities as HD video, face recognition, and cloud workloads. Project Sumatra is an OpenJDK project targeted at bringing Java to such heterogeneous platforms--with hardware and software experts working together to modify the JVM for these advanced applications and platforms.Ramani next discussed the latest with Java in the embedded space--"the Internet of things" and M2M--declaring this to be "the next IT revolution," with Java as the ideal technology for the ecosystem. Last week, Oracle released Java ME Embedded 3.2 (for micro-contollers and low-power devices), and Java Embedded Suite 7.0 (a middleware stack based on Java SE 7). Axel Hansmann, VP Strategy and Marketing, Cinterion, explored his company's use of Java in M2M, and their new release of EHS5, the world's smallest 3G-capable M2M module, running Java ME Embedded. Hansmaan explained that Java offers them the ability to create a "simple to use, scalable, coherent, end-to-end layer" for such diverse edge devices.Marc Brule, Chief Financial Office, Royal Canadian Mint, also explored the fascinating use-case of JavaCard in his country's MintChip e-cash technology--deployable on smartphones, USB device, computer, tablet, or cloud. In parting, Ramani encouraged developers to download the latest releases of Java Embedded, and try them out.Cameron Purdy, VP, Fusion Middleware Development and Java EE, summarized the latest developments and announcements in the Enterprise space--greater developer productivity in Java EE6 (with more on the way in EE 7), portability between platforms, vendors, and even cloud-to-cloud portability. The earliest version of the Java EE 7 SDK is now available for download--in GlassFish 4--with WebSocket support, better JSON support, and more. The final release is scheduled for April of 2013. Nicole Otto, Senior Director, Consumer Digital Technology, Nike, explored her company's Java technology driven enterprise ecosystem for all things sports, including the NikeFuel accelerometer wrist band. Looking beyond Java EE 7, Purdy mentioned NoSQL database functionality for EE 8, the concurrency utilities (possibly in EE 7), some of the Avatar projects in EE 7, some in EE 8, multi-tenancy for the cloud, supporting SaaS applications, and more.Rizvi ended by introducing Dr. Robert Ballard, oceanographer and National Geographic Explorer in Residence--part of Oracle's philanthropic relationship with the National Geographic Society to fund K-12 education around ocean science and conservation. Ballard is best known for having discovered the wreckage of the Titanic. He offered a fascinating video and overview of the cutting edge technology used in such deep-sea explorations, noting that in his early days, high-bandwidth exploration meant that you’d go down in a submarine and "stick your face up against the window." Now, it's a remotely operated, technology telepresence--"I think of my Hercules vehicle as my equivalent of a Na'vi. When I go beneath the sea, I actually send my spirit." Using high bandwidth satellite links, such amazing explorations can now occur via smartphone, laptop, or whatever platform. Ballard’s team regularly offers live feeds and programming out to schools and the world, spanning 188 countries--with embedding educators as part of the expeditions. It's technology at its finest, inspiring the next-generation of scientists and explorers!

    Read the article

  • Hologic Ensures Regulatory Compliance & UDI with Agile PLM for the Medical Device Industry

    - by Ulf Köster
    A new success story featuring Hologic, Inc., is now available. Hologic is known for developing innovative medical technology—like the world’s first 3-D mammogram—that can quickly diagnose women’s health issues and save lives in the process.The success story features Hologic’s use of Oracle Agile PLM to ensure regulatory compliance in every phase of product development, including managing all product-related data, design history files, and device master records. Hologic is using Oracle Agile PLM as the foundation for Unique Device Identification (UDI). Thanks to Agile PLM, Hologic can easily interface with the FDA’s database (GUDID) to streamline compliance, without devoting additional time and resources towards a new solution. Hologic is one of the first 2 companies granted production accounts by the FDA for GUDID submittal, and is the first company to submit official data. This an important milestone for Oracle Agile PLM, our partner Inspirage and the Medical Device industry as a whole. Read the full story here!

    Read the article

  • Google I/O 2010 - Porting v2 JavaScript Maps API apps to v3

    Google I/O 2010 - Porting v2 JavaScript Maps API apps to v3 Google I/O 2010 - Stepping up: Porting v2 JavaScript Maps API applications to v3 Geo 201 Daniels Lee The JavaScript Maps API v3 is the future of the Google Maps API. To take advantage of the many great features coming to the API you will need to migrate existing v2 applications to v3. This session will guide you through the process, illustrating how easy it is to start reaping the benefits in features and performance. For all I/O 2010 sessions, please go to code.google.com From: GoogleDevelopers Views: 10 0 ratings Time: 01:04:07 More in Science & Technology

    Read the article

  • JCP Calendar for 2013 - First EC Meeting 15-16 January

    - by Heather VanCura
    The JCP 2013 calendar and EC Meeting schedule has now been finalized and published :-). This year the EC will be holding meetings in the San Francisco Bay Area in January, and also in September, scheduled around the JavaOne San Francisco Conference; and, Credit Suisse will host the May EC Meeting in Zurich, Switzerland.  The first JCP EC Meeting of 2013 will be a Face to Face Meeting in Santa Clara, California USA, hosted by Intel.  We are of in the midst preparing the agenda now, but it will include a 2012 annual review, JSR Spec Lead presentations & updates, as well as JSR 358, A major revision of the Java Community Process, Expert Group session. You can view meeting materials and minutes from the JCP EC Meetings on JCP.org. The JCP also plans to meet with the Java User Group Leaders attending the User Group Leaders Summit being held at Oracle Redwood Shores location 14-16 January.

    Read the article

  • How To Colorize Black and White Vintage Photographs in Photoshop

    - by Eric Z Goodnight
    Ever wanted to add color to your old, vintage, or historical photographs? Load up some old pictures and see how color can be added quickly to any black and white photograph in this simple Photoshop how to. While many purists simply don’t like the look of colorized black and white photographs, the ability to add color to black and white images is as indispensible as it is simple. Read on to see just how easy it can be Latest Features How-To Geek ETC How To Create Your Own Custom ASCII Art from Any Image How To Process Camera Raw Without Paying for Adobe Photoshop How Do You Block Annoying Text Message (SMS) Spam? How to Use and Master the Notoriously Difficult Pen Tool in Photoshop HTG Explains: What Are the Differences Between All Those Audio Formats? How To Use Layer Masks and Vector Masks to Remove Complex Backgrounds in Photoshop Hack a Wireless Doorbell into a Snail Mail Indicator Enjoy Clutter-Free YouTube Video Viewing in Opera with CleanTube Bring Summer Back to Your Desktop with the LandscapeTheme for Chrome and Iron The Prospector – Home Dash Extension Creates a Whole New Browsing Experience in Firefox KinEmote Links Kinect to Windows Why Nobody Reads Web Site Privacy Policies [Infographic]

    Read the article

  • High Speed Photographs Capture Pellet Gun Destruction

    - by Jason Fitzpatrick
    What do you get when you combine high speed flash photography, a carefully focused camera, and pellet gun? Gloriously detailed pictures of pellets tearing apart fruit, cans, ceramic gnomes, and more. Alan Sailer has a passion; in his garage studio he photographs all manner of objects–bottles, raspberries, candy, soda cans–at the moment a pellet shot from a pellet gun tears them apart. The results are beautiful and reminescent of early high-speed photos by photography pioneer Edgerton Born. Hit up the link below to check out the collection and read more about his process. Alan Sailer’s High Speed Photographs [via FlavorWire] Make Your Own Windows 8 Start Button with Zero Memory Usage Reader Request: How To Repair Blurry Photos HTG Explains: What Can You Find in an Email Header?

    Read the article

  • Annotation Processing Virtual Mini-Track at JavaOne 2012

    - by darcy
    Putting together the list of JavaOne talks I'm interested in attending, I noticed there is a virtual mini-track on annotation processing and related technology this year, with a combination of bofs, sessions, and a hands-on-lab: Monday Multidevice Content Display and a Smart Use of Annotation Processing, Dimitri BAELI and Gilles Di Guglielmo Tuesday Advanced Annotation Processing with JSR 269, Jaroslav Tulach Build Your Own Type System for Fun and Profit, Werner Dietl and Michael Ernst Wednesday Annotations and Annotation Processing: What’s New in JDK 8?, Joel Borggrén-Franck Thursday Hack into Your Compiler!, Jaroslav Tulach Writing Annotation Processors to Aid Your Development Process, Ian Robertson As the lead engineer on bot apt (rest in peace) in JDK 5 and JSR 269 in JDK 6, I'd be heartened to see greater adoption and use of annotation processing by Java developers.

    Read the article

  • Managing the Transition to IFRS

    As countries around the world announce and begin their move to adopting IFRS what can companies learn from those that have already travelled this path? Nigel Youell, Product Marketing Director for Performance Management Applications at Oracle talks to David Jones, Director at PWC, who has worked with multi-national companies across Europe helping them to make this transition and to improve their financial reporting in the process. This podcast offers those who have not yet started, or are currently undertaking, the IFRS journey the chance to learn from David's considerable experience on how to make IFRS an opportunity for improvement rather than just an enforced change.

    Read the article

< Previous Page | 458 459 460 461 462 463 464 465 466 467 468 469  | Next Page >