Search Results

Search found 13575 results on 543 pages for 'desktop virtualization'.

Page 210/543 | < Previous Page | 206 207 208 209 210 211 212 213 214 215 216 217  | Next Page >

  • Java JRE 7 Certified with Oracle E-Business Suite

    - by LuciaC
    Java Runtime Environment 7u10 (a.k.a. JRE 1.7.0_10 build 18) and later updates on the JRE 7 codeline are now certified with Oracle E-Business Suite Release 11i and 12 Windows-based desktop clients. What do you need to do and where are the official patch requirements? EBS customers should ensure that they are running JRE 7u10, at minimum, on Windows desktop clients.  There are also official patch requirements to avoid compatibility issues of E-Business Suite with JRE 7.  Please refer to Steven Chan's Oracle E-Business Suite Technology blog for all the detailed information regarding this certification.

    Read the article

  • Ubuntu One not syncing fully

    - by wurlyfan
    I have uploaded several folders of data to Ubuntu One from my desktop computer, over the last few weeks, and I can see that all the contents are there when I look at my account on the web. When I look at my laptop (connected to the same account), one of the first folders I uploaded hasn't downloaded completely. New folders added from either device seem to sync correctly, but this one older folder remains almost empty, even though the control panel says file syncing is up-to-date. I have plenty of space available. Stopping and restarting the sync daemon and rebooting the laptop are both ineffective. What can I do to make this folder sync fully? I don't want to risk losing the data (which now exists only in Ubuntu One), and I don't have a lot of broadband data to play with. I've seen several bugs relating to this sort of issue but they're all quite old and apparently fixed, while this is happening on new 13.04 installations on both desktop and laptop.

    Read the article

  • Setting umask globally

    - by DevSolar
    I am using a private user group setup, i.e. a user foo's home directory is owned by foo:foo, not foo:users. For this to work, I need to set the umask to 002 globally. After a quick grep -RIi umask /etc/*, it seemed for a moment that modifying the UMASK entry in /etc/login.defs should do the trick. It does, too -- but only for console logins. If I log in to my desktop, and open a terminal there, I still get to see the default umask 022. Same goes for files created from apps started through the menu. Apparently, the display manager (or whatever X11 component responsible) does source some different setting than a console login does, and damned if I could tell which one it is. (I tried changing the setting in /etc/init.d/rc, and no, it did not help.) How / where do I set umask globally, so that the X11 desktop environment gets the memo as well?

    Read the article

  • 12.04 LTS won't install from CD

    - by Rob Hays
    I've been trying to install 12.04 LTS onto a Dell with a PIII from CD. Booting from the CD the install gets through the "Who are you" process, begins copying files. The progress bar gets as far as the last period in "Copying files...". The box clears, and an error box comes up "The installer has encountered an unrecoverable error. A desktop session will now be run so that you may investigae the problem or try installing again." When I try to install from this desktop session, the install gets to the same point, the copying files box closes, and then just stops. The pointer is busy, the cd drive spins up occaisonally with no data transfer, no hard drive activity. When I boot from the CD and access the disk boot menu, the disk checks good, memory checks good ( I upgraded the original memory to 512 mb). I also updated the bios to the newest from Dell. This is an older L866r, but should meet the requirements.

    Read the article

  • How do I select DWM or fluxbox to start on login?

    - by axe
    I am trying to setup DWM in Ubuntu 10.10. The package is successfully built - and now I am trying to get it up and running. I created the dwm.desktop file with the proper text and the path to the dwm exec is correct. On login screen, I can see dwm as a alternate session : but if I select it, still gnome starts. I even tried another window manager (fluxbox) - same problem. I can see and select it in the login screen, but gnome starts up. Trying gconftool-2 -s /desktop/gnome/session/required_components/windowmanager dwm --type string also does nothing. Killing the windows manager just restarts it - even if I chvt first. Also, don't see the xinit daemon running. Any ideas ? This setup was trivial in Debian, and older versions of Ubuntu. What changed ? I am guessing gnome integration is to blame - but am clueless as of now.

    Read the article

  • SproutCore : un framework JavaScript pour enrichir les interfaces web et les rendre semblables à des

    SproutCore : un framework JavaScript pour enrichir les UI Et les rendre semblables à celles d'applications desktop SproutCore est un framework JavaScript encore confidentiel, mais qui commence à faire parler de lui. Edité par le société Sproutit, cette technologie s'adresse aux développeurs webs. Elle leur permet d'enrichir les interfaces utilisateurs pour les rendre quasi-semblables à celles d'applications desktop. «*AJAX a été utilisé pour créer des pages web qui peuvent se mettre à jour sans avoir à se recharger dans le navigateur. Mais elles ressemblent toujours à des pages webs et elles sont généralement limitées au niveau de l'interactivité » peut-on lire sur ...

    Read the article

  • DesktopCouch does not sync

    - by burli
    I have here two computers, both installed with Ubuntu 10.10. The Laptop uses the Netbook Edition. Both are connected, file sync and Tomboy Notes are working. My Desktop syncs databases with Ubuntu One, my Laptop does not. If I delete for example the contacts Database on my Desktop, the database is back a view seconds later. If I delete the same database on my laptop, nothing happens. (I'm using Futon) If I kill beam, delete the config and try to restart with the dbus-send I got an error Error org.freedesktop.DBus.Error.NoReply: Did not receive a reply. Possible causes include: the remote application did not send a reply, the message bus security policy blocked the reply, the reply timeout expired, or the network connection was broken. How can I fix that?

    Read the article

  • How do I remove nitrogen back to Ubuntu's default wallpaper manager?

    - by jonalmeida
    I was using nitrogen when I using openbox, but now I'm trying to stop it from setting the wallpaper so I can use the Ubuntu default (i.e via the System Settings panel). What I've done so far: sudo apt-get remove nitrogen Removed nitrogen configs at ~/.config/nitrogen/ Checked to make sure that draw_background in /desktop/gnome/background was checked So far it still doesn't work. I can't right-click on the desktop, so I'm guessing that I've missed something to get Nautilus back to drawing icons on the screen. Any help is much appreciated. Thanks!

    Read the article

  • Can't see YouTube videos in either Firefox or Chromium

    - by williepabon
    Dont' know why or if it is related, after upgrading Ubuntu 10.04 kernel ( I guess from 2.6.32-38 to -42) I can's see youtube videos. I checked the flash plug-ins installed and I have: (1) flashplugin-installer, (2) flashplugin-nonfree and (3) flashplugin-nonfree-extrasound. I haven't reinstalled the browsers yet. Here's the code of the plugins installed: williepabon@raquel-desktop:~$ sudo lsb_release -a; uname -a; dpkg -l | egrep 'flash|gnash|swf|spark' [sudo] password for williepabon: No LSB modules are available. Distributor ID: Ubuntu Description: Ubuntu 10.04.4 LTS Release: 10.04 Codename: lucid Linux raquel-desktop 2.6.32-42-generic #95-Ubuntu SMP Wed Jul 25 15:57:54 UTC 2012 i686 GNU/Linux ii flashplugin-installer 11.2.202.238ubuntu0.10.04.1 Adobe Flash Player plugin installer ii flashplugin-nonfree 11.2.202.238ubuntu0.10.04.1 Adobe Flash Player plugin installer (transit ii flashplugin-nonfree-extrasound 0.0.svn2431-3 Adobe Flash Player platform support library Any ideas to solve this issue? Thanks.

    Read the article

  • Oracle's Sun x86 Server Product Launch Webcast, April 10th

    - by Larry Wake
    On April 10th, 2012, Oracle will host a webcast to discuss its new generation of x86 servers. Register today Topics covered will include: Enhanced virtualization for consolidation and improved server utilization Reduced licensing costs with 0.5 core factors for Oracle per core-priced software Unparalleled reliability and availability for enterprise environments Increased visibility and efficiency with Oracle Enterprise Manager Ops Center and expert 24/7 support Ongoing protection for your existing software and training investments Live Webcast:The Industry's Best x86 Platform for Running Oracle Enterprise Applications Tuesday, April 10, 2012 9:00 AM PDT 40 minutes including Q&A 

    Read the article

  • Good, simple reasons for having multiple environments

    - by smp7d
    Throughout my career I had worked at companies that had a collection of different environments for different purposes. We always had more or less our desktop environment, a test environment, a QA environment, a staging environment and a production environment. This went for both servers/applications and any data sources we were using. When I started at my current company I found that 90% of the apps were either developed on a desktop environment against production data sources or developed directly on the production server depending on the platform. I wasn't fazed because I was hired in part to make changes to improve the way the development team functioned, which was clear from my interview process. We slowly started to turn the philosophy and pretty soon, most of the apps could be run in either a desktop, test or production environment. Not too long after that staging came around as well. Now most of our developers see the benefit of this methodology and defend it vigilantly. However, we have a number of legacy apps that never got migrated. We also have a number of legacy programmers who think of this as a waste of time. Unfortunately, we got lip service but never full buy-in from management. We got what we thought was a commitment to invest substantially in this about a year ago, but nothing materialized despite the considerable planning that we put into it. Now we are finding that we need more and more environments. We need help from the server/network administration teams for setup and we need participation from the business stakeholders to support the release cycle. We are at a place now where a project can function what I consider "normally" only if you have the right people on the project and the time to set up the proper environments. I'd love to present a complete argument, but management really has no time and interest in hearing me out until there is a critical issue. I can't really articulate the benefits simply as it always just seemed second nature to me. I was wondering if there are any good, simple, irrefutable reasons for the separation of environments that would get managers with no development experience to get behind this idea. Are there any good resources/literature on the topic?

    Read the article

  • How to make a launcher that will first navigate to a folder and then execute a command that resides in normal /usr/bin/

    - by Nirmik
    Okay this Question is basically directed for using GRIVE the linux client for Google Drive Details on how to do it are Here. The thing is that,evrytime i want the folder to sync,I have to navigate to the google drive folder and then execute the grive commnd. I want to make it simple..I want to make a launcher(I know how to make a *.desktop file). But in a .desktop file you always give path to executable file(generally .sh). Here,there is no script in the Grive folder.The app is as usual in /usr/bin/grive Now how do I make the launcher to first navigate to the grive folder and then execut the grive command.. Thanx :)

    Read the article

  • If I create a link to a folder, how can I get from that linked folder to the "real" folder from within Nautilus?

    - by snowguy
    Say I have a folder several layers down in my documents folder. And I want easy access to it from my desktop. To do that I: Go to the parent folder in Nautilus. Right click on the folder's Icon and choose Make Link Cut / Paste the new "Link to ..." folder onto my desktop. Great. And mostly this works fine for me. But suppose I want to get to that folder's parent. I can of course get there using the original path--what Nautilus calls the "link path" which I can see in the properties of the folder. But that seems harder than it ought to be. How can I click on the folder and go to the link path directly?

    Read the article

  • Problems running Ubuntu 12.10 in VMWare Player 5

    - by Tiim
    I'm trying to install Ubuntu 12.10 in VMWare Player 5. My pc is running Windows 7. The installation seems to go okay, but when I boot up ubuntu, although it starts up and I see the desktop, I can't see any task bar or basically anything other than the default wallpaper in the background. When I move my cursor across the screen towards the edges, the desktop flickers and extremely distorted and pixelated objects appear momentarily. I've tried uninstalling everything, downloading it all again and reinstalling, but I get the exact same problem. Does anyone recognise this issue? Is there a more stable pairing than Ubuntu 12.10 & WMWare Player 5.0.1, perhaps?

    Read the article

  • Service Layer - how broad should it be, and should it also be used from the local application?

    - by BornToCode
    The background: I need to build a desktop application with some operations (CRUD and more) (=winforms), I need to make another application which will re-use some of the functions of the main application (=webforms). I'm using service layer for reusing my functions. The service is calling the functions on the BL layer (correct me if I'm doing this wrong). so my desktop has 4 projects - DAL, BL, UI, WEBSERVICES. The dilemma (simple but I still need some more experienced opinions): In my main winform UI - should I call the functions from the BL - bl.getcustomers(), or do it similar to how I call it in the webform, and call the functions from the service - webservices.getcustomers? Should I create a service for every single function on the BL even if I need some of the functions only in one UI? for example - should I create services for all the CRUD operations, even though I need to re-use only update operation in the webform? YOUR HELP IS MUCH APPRECIATED

    Read the article

  • Reasons to Use a VM For Development

    - by George Stocker
    Background: I work at a start-up company, where one team uses Virtual Machines to connect to a remote server to do their development, and another team (the team I'm on) uses local IIS/SQL Server 2005/Visual Studio installations to conduct work. Team VM is located about 1000 miles from Team Non-VM, and the servers the VMs run off of are located near Team VM (Latency, for those that are wondering, is about 50ms). A person high in the company is pushing for Team Non-VM to use virtual machines for programming, development, and testing. The latter point we agree on -- we want Virtual Machines to test configurations and various aspects of the web application in a 'clean' state. The Problem: What we don't agree on is having developers using RDP to connect to a desktop remotely that contains Visual Studio, SQL Server, and IIS to do the same development we could do locally on our laptops. I've tried the VM set-up, and besides the color issue, there is a latency issue that is rather noticeable, not to mention that since we're a start-up, a good number of employees work from home on occasion with our work laptops, and this move would cut off the laptops. They'd be turned in. Reasons to Use Remote VMs for Development (Not Testing!): Here are the stated reasons that this person wants us to use VMs: They work for TeamVM. They keep the source code "safe". If we want to work from home, we could just use our home PCs. Licenses (I don't know what the argument is, only that it's been used). Reasons not to use Remote VMs for Development: Here are the stated reasons why we don't want to use VMs: We like working from home. We get a lot done on our own time. We're not going to use our Home PCs to do work related stuff. The Latency is noticeable. Support for the VMs (if they go down, or if we need a new VM) takes a while. We don't have administrative privileges on the VM, and are unable to change settings as needed. What I'm looking for from the community is this: What reasons would you give for not using VMs for development? Keep in mind these are remote VMs -- this isn't a VM running on a local desktop. It's using the laptop (or a desktop) as a thin client for a remote VM. Also, on the other side of the coin: Is there something we're missing that makes VMs more palatable for development? Edit: I think 'safe' is used in term of corporate espionage, or more correctly if the Laptop gets stolen, the person who stole would have access to our source code. The former (as we've pointed out, is always going to be a possibility -- companies stop that with litigation, there isn't a technical solution (so far as I can see)). The latter point is ( though I don't know its usefulness in a corporate scenario) mitigated by Truecrypt'ing the entire volume.

    Read the article

  • X crash at login for 1 user

    - by marxjohnson
    User switching wasn't working on my 12.04 LTS desktop (just dropped me to TTY8 with a blinking cursor) so I tried to manually start a second X session by logging in to TTY6 and running startx -- :1. This didn't work either, and my machine locked up. Now when I try to log in as the second user from LightDM, X instantly crashes and I'm thrown back to the login screen. Other accounts on the machine work fine, and it happens for every desktop environment. I've had a poke around in my home directory, but I can't see anything obvious to change/delete to get it working again. Can anyone advise please?

    Read the article

  • Different behavior when launched from terminal instead of Unity launcher

    - by dgkontopoulos
    I have written two Perl/Gtk programs. When launched from the dash menu, they run just fine. However, if I try to launch them from terminal using the very same command found in their .desktop files, their Unity launcher will be blurry and will remain inactive when clicked, if I keep it in the Unity bar. The problem is solved if the Exec part of the desktop file is replaced with perl path_to_script However, that leads to other problems, including a lintian warning and forcing all Perl GUI applications running from terminal to adopt the same launcher. This issue is quite annoying since one of the programs relies on a different (Python) program in order to be launched and this results in having a blurry and inactive launcher.

    Read the article

  • Creating Visual Studio Extension Files (VSIX) for Template Deployment

    While working on some plugins for the new Seesmic Desktop PlatformI got sick of copying and pasting some boiler plate code over and over. I had created some helper templates for myself so that I could say FileNew Seesmic Desktop Plugin and get everything I needed initially. This weekend I had some time and formalized those templates into an easy-to-use installer for anyone to consume. NOTE: It is likely that Seesmic themselves will create developer project/item templatesthese were for my own...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Chromium web-app creation doesn't work in 12.10?

    - by speter
    If you create an application of a website in Chromium and choose "desktop", a .desktop file is created. In 12.04, you were able to move it to ~/,local/share/applications/ and then start it from the Dash or the launcher. In my fresh 12.10 installation, this doesn't work anymore. I think the line that is misinterpreted is Exec=/usr/bin/chromium-browser --app=http://buymeapie.com/. If you try to run it from a terminal, you get the following result: 20:32 ~ speter Exec=/usr/bin/chromium-browser --app=http://buymeapie.com/ bash: --app=http://buymeapie.com/: Datei oder Verzeichnis nicht gefunden (English: "File or directory not found"). Can anyone explain or knows a workaround?

    Read the article

  • FOUR questions to ask if you are implementing DATABASE-AS-A-SERVICE

    - by Sudip Datta
    During my ongoing tenure at Oracle, I have met all types of DBAs. Happy DBAs, unhappy DBAs, proud DBAs, risk-loving DBAs, cautious DBAs. These days, as Database-as-a-Service (DBaaS) becomes more mainstream, I find some complacent DBAs who are basking in their achievement of having implemented DBaaS. Some others, however, are not that happy. They grudgingly complain that they did not have much of a say in the implementation, they simply had to follow what their cloud architects (mostly infrastructure admins) offered them. In most cases it would be a database wrapped inside a VM that would be labeled as “Database as a Service”. In other cases, it would be existing brute-force automation simply exposed in a portal. As much as I think that there is more to DBaaS than those approaches and often get tempted to propose Enterprise Manager 12c, I try to be objective. Neither do I want to dampen the spirit of the happy ones, nor do I want to stoke the pain of the unhappy ones. As I mentioned in my previous post, I don’t deny vanilla automation could be useful. I like virtualization too for what it has helped us accomplish in terms of resource management, but we need to scrutinize its merit on a case-by-case basis and apply it meaningfully. For DBAs who either claim to have implemented DBaaS or are planning to do so, I simply want to provide four key questions to ponder about: 1. Does it make life easier for your end users? Database-as-a-Service can have several types of end users. Junior DBAs, QA Engineers, Developers- each having their own skillset. The objective of DBaaS is to make their life simple, so that they can focus on their core responsibilities without having to worry about additional stuff. For example, if you are a Developer using Oracle Application Express (APEX), you want to deal with schema, objects and PL/SQL code and not with datafiles or listener configuration. If you are a QA Engineer needing database copies for functional testing, you do not want to deal with underlying operating system patching and compliance issues. The question to ask, therefore, is, whether DBaaS makes life easier for those users. It is often convenient to give them VM shells to deal with a la Amazon EC2 IaaS, but is that what they really want? Is it a productive use of a developer's time if he needs to apply RPM errata to his Linux operating system. Asking him to keep the underlying operating system current is like making a guest responsible for a restaurant's decor. 2. Does it make life easier for your administrators? Cloud, in general, is supposed to free administrators from attending to mundane tasks like provisioning services for every single end user request. It is supposed to enable a readily consumable platform and enforce standardization in the process. For example, if a Service Catalog exposes DBaaS of specific database versions and configurations, it, by its very nature, enforces certain discipline and standardization within the IT environment. What if, instead of specific database configurations, cloud allowed each end user to create databases of their liking resulting in hundreds of version and patch levels and thousands of individual databases. Therefore the right question to ask is whether the unwanted consequence of DBaaS is OS and database sprawl. And if so, who is responsible for tracking them, backing them up, administering them? Studies have shown that these administrative overheads increase exponentially with new targets, and it could result in a management nightmare. That leads us to our next question. 3. Does it satisfy your Security Officers and Compliance Auditors? Compliance Auditors need to know who did what and when. They also want the cloud platform to be secure, so that end users have little freedom in tampering with it. Dealing with VM sprawl is not the easiest of challenges, let alone dealing with them as they keep getting reconfigured and moved around. This leads to the proverbial needle in the haystack problem, and all it needs is one needle to cause a serious compliance issue in the enterprise. Bottomline is, flexibility and agility should not come at the expense of compliance and it is very important to get the balance right. Can we have security and isolation without creating compliance challenges? Instead of a ‘one size fits all approach’ i.e. OS level isolation, can we think smartly about database isolation or schema based isolation? This is where the appropriate resource modeling needs to be applied. The usual systems management vendors out there with heterogeneous common-denominator approach have compromised on these semantics. If you follow Enterprise Manager’s DBaaS solution, you will see that we have considered different models, not precluding virtualization, for different customer use cases. The judgment to use virtual assemblies versus databases on physical RAC versus Schema-as-a-Service in a single database, should be governed by the need of the applications and not by putting compliance considerations in the backburner. 4. Does it satisfy your CIO? Finally, does it satisfy your higher ups? As the sponsor of cloud initiative, the CIO is expected to lead an IT transformation project, not merely a run-of-the-mill IT operations. Simply virtualizing server resources and delivering them through self-service is a good start, but hardly transformational. CIOs may appreciate the instant benefit from server consolidation, but studies have revealed that the ROI from consolidation would flatten out at 20-25%. The question would be: what next? As we go higher up in the stack, the need to virtualize, segregate and optimize shifts to those layers that are more palpable to the business users. As Sushil Kumar noted in his blog post, " the most important thing to note here is the enterprise private cloud is not just an IT project, rather it is a business initiative to create an IT setup that is more aligned with the needs of today's dynamic and highly competitive business environment." Business users could not care less about infrastructure consolidation or virtualization - they care about business agility and service level assurance. Last but not the least, lot of CIOs get miffed if we ask them to throw away their existing hardware investments for implementing DBaaS. In Oracle, we always emphasize on freedom of choosing a platform; hence Enterprise Manager’s DBaaS solution is platform neutral. It can work on any Operating System (that the agent is certified on) Oracle’s hardware as well as 3rd party hardware. As a parting note, I urge you to remember these 4 questions. Remember that your satisfaction as an implementer lies in the satisfaction of others.

    Read the article

  • Can't login to my XFCE environment from lxdm

    - by Noob
    I installed lxdm as my display manager because I heard its very lightweight. Now after installing it when I select my Desktop as Xfce session I am not able to login. Basically what happens is it waits for sometime and than reload the same login page. On the other hand if I select lxde as my Desktop I am able to login. So, what am I doing wrong here. The other problem I am facing with lxdm is it doesn't show my default user account. I have to select more and enter both my username and password to login into any session.

    Read the article

< Previous Page | 206 207 208 209 210 211 212 213 214 215 216 217  | Next Page >