Search Results

Search found 37240 results on 1490 pages for 'ubuntu enterprise cloud'.

Page 345/1490 | < Previous Page | 341 342 343 344 345 346 347 348 349 350 351 352  | Next Page >

  • How can I install ATI Catalyst Video driver of Ubuntu 12.04.3?

    - by Eastsun
    EDIT My computer is Thinkpad E420 with Intel HD3000 and ATI 6630m. I have reinstalled my Ubuntu12.04.3 yesterday. I had successfully installed Ubuntu and ATI Catalyst Video driver many times about half year ago following these links: How do I get AMD/Intel Hybrid Graphics drivers to work? What is the correct way to install ATI Catalyst Video Drivers (fglrx)? However, this time I have no luck with installing ATI Catalyst Video driver. I have tried version 13.4, 13.1, 12.8 of ATI Catalyst Video driver. It seemed that 13.4 doesn't support 12.04 any more, following this link , the lowest Ubuntu version supported by ATI 13.4 is 12.10 and for the reset of the ATI drivers, the kernel version and xorg version of Ubuntu 12.04.3 is too high to be supported. So, should I stay on ubuntu 12.04.1 with ATI 12.8, or there is any way to install ATI driver on Ubuntu 12.04.3?

    Read the article

  • Can not install Ubuntu 12.04 or 12.10 on Toshiba qosmio x870. Please help!

    - by Mighty
    I have a new Toshiba qosmio x870 and for the past one week I have been trying to install Ubuntu 12.04 from a USB and Live CD without success. I keep on getting this error: Boot failure: a proper digital signature was not found. One or more files on the selected boot device was rejected by the Secure Boot feature. I even tried installing Ubuntu with the Windows installer. After installation and I reboot the PC, first I see the error that points to: \ubuntu\winboot\wubildr.mbr Status: 0xc000007b Info: The OS couldn't be loaded because a required file is missing or contains errors. When I restart, that the previous error doesn't show up and I see both Windows 8 and Ubuntu (happy that I was successful) but when I click on Ubuntu, it flags an error. This is the first time I'm having a Secure Boot-capable PC. What will be the danger in disabling the secure boot? I'll be happy if I can get assistance from anyone.

    Read the article

  • What video graphic card would permit ubuntu's standard driver to work well?

    - by Rick
    I installed ubuntu 12.04. All seemed well until I installed the nvidia driver. Then crashola! This situation is untenable. It seems I cannot trust nvidia, and it seems that I cannot rely on ubuntu gurus to test 3rd party drivers. So, apparently some video card manufacturers do not care enough about the linux market to test their drivers, or are there too many 3rd party video cards so that ubuntu folks do not test any 3rd party video drivers? Hence the question: What video graphic card would permit me to use ubuntu's standard driver so that I do not have to rely on nvidia's or any other 3rd-party driver? Perhaps I could then install THAT card and have things work? The ubuntu standard driver actually worked prior to installing the nvidia driver, but not well, and that was because the display flickered, and flickering gives me a headache.

    Read the article

  • Git-Based Source Control in the Enterprise: Suggested Tools and Practices?

    - by Bob Murphy
    I use git for personal projects and think it's great. It's fast, flexible, powerful, and works great for remote development. But now it's mandated at work and, frankly, we're having problems. Out of the box, git doesn't seem to work well for centralized development in a large (20+ developer) organization with developers of varying abilities and levels of git sophistication - especially compared with other source-control systems like Perforce or Subversion, which are aimed at that kind of environment. (Yes, I know, Linus never intended it for that.) But - for political reasons - we're stuck with git, even if it sucks for what we're trying to do with it. Here are some of the things we're seeing: The GUI tools aren't mature Using the command line tools, it's far to easy to screw up a merge and obliterate someone else's changes It doesn't offer per-user repository permissions beyond global read-only or read-write privileges If you have a permission to ANY part of a repository, you can do that same thing to EVERY part of the repository, so you can't do something like make a small-group tracking branch on the central server that other people can't mess with. Workflows other than "anything goes" or "benevolent dictator" are hard to encourage, let alone enforce It's not clear whether it's better to use a single big repository (which lets everybody mess with everything) or lots of per-component repositories (which make for headaches trying to synchronize versions). With multiple repositories, it's also not clear how to replicate all the sources someone else has by pulling from the central repository, or to do something like get everything as of 4:30 yesterday afternoon. However, I've heard that people are using git successfully in large development organizations. If you're in that situation - or if you generally have tools, tips and tricks for making it easier and more productive to use git in a large organization where some folks are not command line fans - I'd love to hear what you have to suggest. BTW, I've asked a version of this question already on LinkedIn, and got no real answers but lots of "gosh, I'd love to know that too!"

    Read the article

  • Ubuntu installer thinks my drive is empty, does not see windows paritions.

    - by John
    Hello, I am trying to install Ubuntu10.10 64bit. I just installed Windows 7 64bit moments ago. My drive is currently partitioned like so: 100MB Boot partition (automatically made by windows 7 installer) 390GB Windows partition ~1.6TB free space When I go through the Ubuntu installer it does not give me the option to install alongside another operating system. My only options are to use the entire disk or to specify partitions manually. When I chose to specify partitions manually it tells me that the drive is all free space! Windows is still booting and behaving normally, and I had not doing anything in Windows yet (had simply installed, booted for first time, then immediately restarted). I am even able to mount the windows partition within Ubuntu Live CD, and see it in the disk viewer (not gparted). Gparted in Ubuntu Live CD again reports no partitions, all free space. Not sure what to do :S. I have installed Windows7 alongside Ubuntu countless times, even Ubuntu 10.10. Thank you very much for your help :).

    Read the article

  • Is there anywhere where I can get Ubuntu installed for me?

    - by user3911336
    I have tried for days to install Ubuntu on my lenovo ideapad z510 and I feel like giving up. I really feel like I will like Ubuntu, but for some reason I always have some king of problem. I really don't want to give up on Ubuntu, but I don't want to brick my laptop either. Is there anywhere I can have Ubuntu installed for me? I have tried messing with my bios, used a DVD and a USB, but no luck. Thanks in advance! Edit: I am based in Warwick, RI in the USA. I don't want to run Ubuntu in windows.

    Read the article

  • How do I install Ubuntu to a USB key?

    - by badp
    If you are hurrying to reply, System ? Administration ? StartUp Disk Creator -- no, that's not what I'm talking about. I want to try Ubuntu 11.04's Unity without touching my existing Ubuntu install. To do this, I need to install the nVidia drivers first (sigh). To do this, I need changes to persist a reboot. To do this, I need to really install Ubuntu on a USB key. How do you do that? What I tried I tried to make a USB key from Testdrive, then boot from it, then choose "Install Ubuntu." The installer refused to install to the installation media itself. I tried, from my installed copy of Ubuntu: sudo kvm /dev/sdb --cdrom .cache/testdrive/iso/ubuntu_natty-desktop-i386.iso ...but the installer didn't detect the disk properly.

    Read the article

  • Ubuntu 12.04 LTS : la beta 2 de "Precise Pangolin" améliore HUD, Ubuntu One et intègre de nouvelles versions d'applications

    Ubuntu 12.04 LTS : la beta 2 de "Precise Pangolin" améliore HUD Ubuntu One et intègre de nouvelles versions d'applications Mise à jour du 09/04/2012 À quelques semaines de la publication de la version finale de Precise Pangolin, la prochaine version du système d'exploitation fondé sur Linux, Canonical publie la dernière beta de l'OS. La beta 2 d'Ubuntu 12.04 LTS apporte des corrections de bugs et quelques nouvelles fonctionnalités, dont l'ajout du noyau Linux 3.2.0-20.33 qui est basé sur la version stable 3.2.12 du Kernel. HUD, le nouveau menu intelligent pour les applications Unity dont un premier aperç...

    Read the article

  • Can't Boot Ubuntu, video drivers or x-server problem?

    - by ZacharyH
    I was uninstalling some programs that I installed to try and get my iPod touch working with Ubuntu (I gave up on that) when ubuntu just crashed. Now after I choose ubuntu in GRUB, it gives me a screen that says "Ubuntu is running in low-graphics mode: your screen, graphics card, and input device settings could not be detected correctly. You will need to configure these yourself" It was working just fine before I started to uninstall those programs. I think that I might have uninstalled something necessary to the system. If I click OK on the screen, it gives me options to reconfigure, troubleshoot, exit to console, or restart X. But no matter what I choose I still can't boot into ubuntu - I get stuck looking at the splash screen which stalls forever. I was receiving support from one of my mate's and he was doing something with the LiveCD, and now the message doesn't pop up any more, I just get stuck at a never ending splash screen. Any help would be appreciated, thanks!

    Read the article

  • Where to install bootloader when installing Ubuntu as secondary OS?

    - by HelpNeeder
    I'm trying to install Ubuntu as secondary OS on my laptop. I have Windows 8 already installed on my laptop. Now, I know how to run Ubuntu from USB drive, I created addition partition and formatted it to EXT4. So I'm ready to install. Now, 'Device for boot loader installation:' displays: /dev/sta ATA HITACHI (750 GB) /dev/sta1 Windows 8 (loader) /dev/sta2 /dev/sta5 /dev/sta6 Ubuntu 12.04 (12.04) /dev/stb I tries choosing Ubuntu 12.04 partition but it doesn't even let me to pick which OS to install and goes straight to Windows 8. Which partition I must choose to be able to pick which OS to boot from? Preferably, set up so Windows 8 will be at first place, and Ubuntu on second. Any ideas? I don't want to mess up anything if I pick something wrong.

    Read the article

  • Enterprise Library--Validator.cs How Abstract Class having definition??

    - by Soham
    Consider this piece of code: public abstract class Validator { protected Validator() { } protected abstract void ValidateCore(object instance, string value, IList<ValidationResult> results); public void Validate(object instance, string value, IList<ValidationResult> results) { if (null == instance) throw new ArgumentNullException("instance"); if (null == results) throw new ArgumentNullException("results"); ValidateCore(instance, value, results); } } TAKE a look at Validate() overload, how can an abstract class have definitions like this?

    Read the article

  • How does thumbnail preview in Ubuntu differ from that of Windows? [closed]

    - by Forbidden Overseer
    Possible Duplicate: How does Ubuntu know what file type a file without extension is I thought this question might get a better response in AskUbuntu, as it seems to have more to do with Ubuntu than Windows at a glance. Let's say I have a foo.mkv file. Thumbnail previews work in both Windows 7 and Ubuntu. When I change the filename to anything random like foo.bar or when I remove the extension itself (making it just foo), Nautilus shows thumbnails normally like if it can recognize what type of files they are - without looking at file extension. This however, doesn't happen in Windows 7. Windows starts asking me things like which application I want to use to open that file as soon as I remove file extension (forget thumbnails...) etc. So, How does this thumbnail preview work in Windows 7 and Ubuntu? What makes Ubuntu recognize files "out of the box" unlike Windows 7?

    Read the article

  • How can I implement full disk encryption for a disk containing Ubuntu with plausible deniability?

    - by Rupert
    I would like to have a disk that: has Ubuntu installed is fully encrypted is setup in such a way as to make denial of the existence of the Ubuntu install plausible Truecrypt provides the last two features but only for Windows: http://www.truecrypt.org/docs/?s=sys-encryption-supported-os The alternate installer for Ubuntu provides the first two features but not the last. I imagine that plausible deniability would be achieved by, at least: Having two installs of Ubuntu and 2 passphrases (or keyfiles) which would unencrypt each one. Moving any unencrypted data (such as /boot) onto a USB stick Are there any programs out there which support this feature set for Ubuntu?

    Read the article

  • How do I get Ubuntu One on Win7 to authenticate my login?

    - by Fred jones
    I just got a new computer running Windows 7 home premium, I used to have Ubuntu One working great on my Ubuntu desktop, but now, running win7, I know my login email address and password is correct, but it still says 'Authentication failed'. Googling the problem, looks like it may be because a device was removed from my list of devices, but the only device listed is my previous Ubuntu workstation, and nothing has been removed. I also checked the windows firewall and Ubuntu One is allowed on all interfaces on incoming and outgoing, and still authentication fails. Web login to my Ubuntu One account works fine.

    Read the article

  • How do you hibernate/suspend Ubuntu like Windows or both at the same time?

    - by All Star Geek
    I'm using an ASUS laptop with Windows 7, and I just installed Ubuntu alongside Windows yesterday, using the no-risk option. I noticed, though, that I can't hibernate Ubuntu the way I did with Windows to save time, and, every time I use Ubuntu, I have to shut down to go into Windows, which has a slow boot. Is there any way that I could have Windows and Ubuntu hibernated at the same time so I can access both easily (preferably without having to hit a function key within seconds). When I try to hibernate or suspend Ubuntu with Windows shut down, the screen just turns blue and doesn't shut the computer off like a Windows hibernate or sleep.

    Read the article

  • I Can't install Skype on Ubuntu 12.04 x64. Can someone help me?

    - by Matheus
    I'm having a trouble when trying to install Skype on my Ubuntu 12.04.1 x64. My notebook have compatible hardware with Ubuntu. My previous OS installed on this machine was the Ubuntu 12.04 x64 and Skype was working great, but I got a formatation and I installed the New ubuntu 12.04.1. When I try to install by the Ubuntu software Center, or by the sudo apt-get install skype, I get this error message (in Portuguese) : "Os pacotes a seguir têm dependências desencontradas: skype: Depend: skype-bin E: impossivel corrigir problemas, você manteve (hold) pacotes quebrados." Something like i have the skype-bin package missing (or broken). I can't find it on Software center. I tried to install it by downloading a deb package with the skype-bin package but it ask for a lot another packages! I tried install the .deb package from the official Skype site, but I got the same error! Any Help? Sorry about my bad English, I'm Brazilian

    Read the article

  • Ubuntu Australia LoCo Team

    - by benonsoftware
    Well first I would like to add at the moment I have had 1174 page views. That does not count people who just browse via the homepage so add around another 200 visitors! Well back on the subject I have recently joined Ubuntu-AU (Australia) LoCo team and it is great! Basilar it is a Ubuntu loco community group and just before the my first time being at a meeting I was emailing the admin and he said that there was a stop for a team reporter and I put my hand up and he said 'yes'. He was actually going to ask at the team meeting I would have said no way! But because I saw it before hand and I asked him I wanted to do it. If you live in Australia consider join Ubuntu-AU they also have a Launchpad team.

    Read the article

  • Can I properly use NS2 on latest Ubuntu 12.04?

    - by Nwe Nwe
    I am a novice user for Ubuntu because I need to work with NS2. NS2 properly work only in Linux, Ubuntu. Now, I wonder if it can properly work on latest version Ubuntu. My current window is Window7. I also want to use Window7. So, which one will be better on parallel booting of Window 7 and Ubuntu or using VMWare on Window7 to install Ubuntu. My RAM is 4G. My processor speed is 1.6G. If I use VMWare, my C drive is only extra 10G. So, does VMWare can work properly in D drive? Please let me know which way is the best for my situation. I hope the advice from anyone. Thanks in advance.

    Read the article

  • Ubuntu 12.10 brise la barrière entre PC et Web, Quantal Quetzal sort avec sa déclinaison serveur

    Ubuntu 12.10 brise la barrière entre PC et Web Quantal Quetzal sort avec Web Apps, Dash et bien plus, la version serveur de l'OS disponible Quatre mois après avoir pointé le bout de son nez en version Alpha, Ubuntu 12.10 « Quantal Quetzal » est prêt pour une utilisation par le grand public. Ubuntu 12.10 marque une étape importante de la symbiose entre Cloud, Web et Desktop entreprise par Canonical. Cette mouture fournit une intégration étroite entre les environnements de bureau, les applications Web et les plateformes de Cloud. La nouvelle fonctionnalité Web Apps du système d'exploitation offre une intégration dans le bureau des applications Web comme Twitter, Gmail ou encore Face...

    Read the article

  • In setting up dual Boot with Windows XP and Ubuntu, which OS do I install first?

    - by markl
    I'd like to install both Ubuntu 12.04 and Windows XP on a Dell laptop, and I was thinking about using a dual boot structure, and using the bulk of my hard drive as empty hard drive space to share files between the two operating systems (so choice of file system type is very important in this set-up). The kind of partitioning structure I would like to use is Partition 1 - Ubuntu 12.04 (root) (20GB) Partition 2 - Ubuntu /home (20GB) Partition 3 : Free Space (560GB) Partition 4 : Windows XP (35GB) Partition 5 : SWAP (3GB) (Total Hardrive Capacity is ~640GB) My question is; what is the best way to go about setting up this kind this system? Should I install Windows XP first and setup the partitions, and then install Ubuntu which I believe will install the GRUB bootloader for OS booting choice or Do I install Ubuntu first, setting up the available partitions and then perform a WIndows install? Please let me know if there is anything in this setup that I have left out and should know about, including things related to setting particular partitions as logical or primary, and whether the boot partition and the filesystem partition should actually be two separate partitions.

    Read the article

  • Dropbox 99% instalation crash. Strange "Searching" Ubuntu Software Center bug [closed]

    - by kocios
    Possible Duplicate: Dropbox fails to install, stuck at 99% I just installed Dropbox via Ubuntu Software Center, and the known crash of 99%, which was in Ubuntu Software center, happened. I want to remove Dropbox completely since something went wrong, but I can't. When I start my Ubuntu, it starts a process which eats up all of the CPU. In Ubuntu Software Center, there is a process "Searching Applying Changes", but nothing really happens. When I'm trying to uninstall Dropbox, nothing happens, since the seaching thing is "first". Please, I'm asking for a option which doesn't include proper reinstallation of Dropbox, only of getting rid of it, and without reinstalling the whole Ubuntu system (format, chaos etc).

    Read the article

  • In the Enterprise Library, how does the abstract Validator.cs have a method definition?

    - by Soham
    Consider this piece of code: public abstract class Validator { protected Validator() { } protected abstract void ValidateCore(object instance, string value, IList<ValidationResult> results); public void Validate(object instance, string value, IList<ValidationResult> results) { if (null == instance) throw new ArgumentNullException("instance"); if (null == results) throw new ArgumentNullException("results"); ValidateCore(instance, value, results); } } Look at the Validate() overload, how can an abstract class have definitions like this?

    Read the article

  • Chroot within chroot

    - by Andy
    I'm using Centos 5.2 and when I try to make a chroot jail using the script, I get: Copying libraries for /usr/bin/scp. (0x00007fff17bfe000) cp: cannot stat `(0x00007fff17bfe000)': No such file or directory ... I am currently using on a rackspace cloud server so i suspect that these dependencies are outside of my own root. Does anyone have a better idea for jailing the sftp server on a cloud server using Centos 5.2?

    Read the article

  • From HttpRuntime.Cache to Windows Azure Caching (Preview)

    - by Jeff
    I don’t know about you, but the announcement of Windows Azure Caching (Preview) (yes, the parentheses are apparently part of the interim name) made me a lot more excited about using Azure. Why? Because one of the great performance tricks of any Web app is to cache frequently used data in memory, so it doesn’t have to hit the database, a service, or whatever. When you run your Web app on one box, HttpRuntime.Cache is a sweet and stupid-simple solution. Somewhere in the data fetching pieces of your app, you can see if an object is available in cache, and return that instead of hitting the data store. I did this quite a bit in POP Forums, and it dramatically cuts down on the database chatter. The problem is that it falls apart if you run the app on many servers, in a Web farm, where one server may initiate a change to that data, and the others will have no knowledge of the change, making it stale. Of course, if you have the infrastructure to do so, you can use something like memcached or AppFabric to do a distributed cache, and achieve the caching flavor you desire. You could do the same thing in Azure before, but it would cost more because you’d need to pay for another role or VM or something to host the cache. Now, you can use a portion of the memory from each instance of a Web role to act as that cache, with no additional cost. That’s huge. So if you’re using a percentage of memory that comes out to 100 MB, and you have three instances running, that’s 300 MB available for caching. For the uninitiated, a Web role in Azure is essentially a VM that runs a Web app (worker roles are the same idea, only without the IIS part). You can spin up many instances of the role, and traffic is load balanced to the various instances. It’s like adding or removing servers to a Web farm all willy-nilly and at your discretion, and it’s what the cloud is all about. I’d say it’s my favorite thing about Windows Azure. The slightly annoying thing about developing for a Web role in Azure is that the local emulator that’s launched by Visual Studio is a little on the slow side. If you’re used to using the built-in Web server, you’re used to building and then alt-tabbing to your browser and refreshing a page. If you’re just changing an MVC view, you’re not even doing the building part. Spinning up the simulated Azure environment is too slow for this, but ideally you want to code your app to use this fantastic distributed cache mechanism. So first off, here’s the link to the page showing how to code using the caching feature. If you’re used to using HttpRuntime.Cache, this should be pretty familiar to you. Let’s say that you want to use the Azure cache preview when you’re running in Azure, but HttpRuntime.Cache if you’re running local, or in a regular IIS server environment. Through the magic of dependency injection, we can get there pretty quickly. First, design an interface to handle the cache insertion, fetching and removal. Mine looks like this: public interface ICacheProvider {     void Add(string key, object item, int duration);     T Get<T>(string key) where T : class;     void Remove(string key); } Now we’ll create two implementations of this interface… one for Azure cache, one for HttpRuntime: public class AzureCacheProvider : ICacheProvider {     public AzureCacheProvider()     {         _cache = new DataCache("default"); // in Microsoft.ApplicationServer.Caching, see how-to      }         private readonly DataCache _cache;     public void Add(string key, object item, int duration)     {         _cache.Add(key, item, new TimeSpan(0, 0, 0, 0, duration));     }     public T Get<T>(string key) where T : class     {         return _cache.Get(key) as T;     }     public void Remove(string key)     {         _cache.Remove(key);     } } public class LocalCacheProvider : ICacheProvider {     public LocalCacheProvider()     {         _cache = HttpRuntime.Cache;     }     private readonly System.Web.Caching.Cache _cache;     public void Add(string key, object item, int duration)     {         _cache.Insert(key, item, null, DateTime.UtcNow.AddMilliseconds(duration), System.Web.Caching.Cache.NoSlidingExpiration);     }     public T Get<T>(string key) where T : class     {         return _cache[key] as T;     }     public void Remove(string key)     {         _cache.Remove(key);     } } Feel free to expand these to use whatever cache features you want. I’m not going to go over dependency injection here, but I assume that if you’re using ASP.NET MVC, you’re using it. Somewhere in your app, you set up the DI container that resolves interfaces to concrete implementations (Ninject call is a “kernel” instead of a container). For this example, I’ll show you how StructureMap does it. It uses a convention based scheme, where if you need to get an instance of IFoo, it looks for a class named Foo. You can also do this mapping explicitly. The initialization of the container looks something like this: ObjectFactory.Initialize(x =>             {                 x.Scan(scan =>                         {                             scan.AssembliesFromApplicationBaseDirectory();                             scan.WithDefaultConventions();                         });                 if (Microsoft.WindowsAzure.ServiceRuntime.RoleEnvironment.IsAvailable)                     x.For<ICacheProvider>().Use<AzureCacheProvider>();                 else                     x.For<ICacheProvider>().Use<LocalCacheProvider>();             }); If you use Ninject or Windsor or something else, that’s OK. Conceptually they’re all about the same. The important part is the conditional statement that checks to see if the app is running in Azure. If it is, it maps ICacheProvider to AzureCacheProvider, otherwise it maps to LocalCacheProvider. Now when a request comes into your MVC app, and the chain of dependency resolution occurs, you can see to it that the right caching code is called. A typical design may have a call stack that goes: Controller –> BusinessLogicClass –> Repository. Let’s say your repository class looks like this: public class MyRepo : IMyRepo {     public MyRepo(ICacheProvider cacheProvider)     {         _context = new MyDataContext();         _cache = cacheProvider;     }     private readonly MyDataContext _context;     private readonly ICacheProvider _cache;     public SomeType Get(int someTypeID)     {         var key = "somename-" + someTypeID;         var cachedObject = _cache.Get<SomeType>(key);         if (cachedObject != null)         {             _context.SomeTypes.Attach(cachedObject);             return cachedObject;         }         var someType = _context.SomeTypes.SingleOrDefault(p => p.SomeTypeID == someTypeID);         _cache.Add(key, someType, 60000);         return someType;     } ... // more stuff to update, delete or whatever, being sure to remove // from cache when you do so  When the DI container gets an instance of the repo, it passes an instance of ICacheProvider to the constructor, which in this case will be whatever implementation was specified when the container was initialized. The Get method first tries to hit the cache, and of course doesn’t care what the underlying implementation is, Azure, HttpRuntime, or otherwise. If it finds the object, it returns it right then. If not, it hits the database (this example is using Entity Framework), and inserts the object into the cache before returning it. The important thing not pictured here is that other methods in the repo class will construct the key for the cached object, in this case “somename-“ plus the ID of the object, and then remove it from cache, in any method that alters or deletes the object. That way, no matter what instance of the role is processing the request, it won’t find the object if it has been made stale, that is, updated or outright deleted, forcing it to attempt to hit the database. So is this good technique? Well, sort of. It depends on how you use it, and what your testing looks like around it. Because of differences in behavior and execution of the two caching providers, for example, you could see some strange errors. For example, I immediately got an error indicating there was no parameterless constructor for an MVC controller, because the DI resolver failed to create instances for the dependencies it had. In reality, the NuGet packaged DI resolver for StructureMap was eating an exception thrown by the Azure components that said my configuration, outlined in that how-to article, was wrong. That error wouldn’t occur when using the HttpRuntime. That’s something a lot of people debate about using different components like that, and how you configure them. I kinda hate XML config files, and like the idea of the code-based approach above, but you should be darn sure that your unit and integration testing can account for the differences.

    Read the article

< Previous Page | 341 342 343 344 345 346 347 348 349 350 351 352  | Next Page >