Search Results

Search found 28744 results on 1150 pages for 'higher order functions'.

Page 488/1150 | < Previous Page | 484 485 486 487 488 489 490 491 492 493 494 495  | Next Page >

  • Ubuntu Screensaver Freezes?

    - by user11833
    I am using ubuntu 10.04. Over the last couple of days, I am having a weird problem: once the machine gets into screen saver, it is impossible to get it "out". I do see the mouse cursor move as I move the mouse, but the login popup (with user / password) does not show up and the screen remains "frozen". Any idea what might be going wrong? For the time being, I have no choice but to restart the computer in order to be able to resume work. Thanks.

    Read the article

  • June 14th Webcast: Using Personalization in Oracle eAM

    - by Oracle_EBS
    ADVISOR WEBCAST: Using Personalization in Oracle eAMPRODUCT FAMILY: Enterprise Asset Management June 14, 2012 at 11 am ET, 9 am MT, 8 am PT Personalization is the ability within an E-Business Suite instance to make changes to the look and behavior of OA Framework-based pages without programming (and, therefore, upgradeable!).TOPICS WILL INCLUDE: Oracle eAM customers have leveraged Personalization to address many of the following use cases: making a non-mandatory field mandatory on a Work Order adding a new field to the header region entering the specifications for a new field disabling a field to prevent user entry adding Asset Description to Create Work Request Page viewing Asset Hierarchy in Create Work Request Page setting Auto Request Material Flag to yes by default A short, live demonstration (only if applicable) and question and answer period will be included. Oracle Advisor Webcasts are dedicated to building your awareness around our products and services. This session does not replace offerings from Oracle Global Support Services. Current Schedule can be found on Note 740966.1 Post Presentation Recordings can be found on Note 740964.1

    Read the article

  • Directory access control with Apache: do I need to use a specific .htaccess?

    - by Mirror51
    I have an Apache webserver, and in the Apache configuration, I have Alias /backups "/backups" <Directory "/backups"> AllowOverride None Options Indexes Order allow,deny Allow from all </Directory> I can access files via http://127.0.0.1/backups. The problem is everyone can access that. I have a web interface, e.g. http://localhost/adminm that is protected with htaccess and password. Now I don't want separate .htaccess and .htpasswd for /backups, and I don't want a second password prompt when a user clicks on /backups in the web interface. Is there any way to use same .htaccess and .htpasswd for the backups directory?

    Read the article

  • Does anyone know of a simple (free?) feature request tracking system we could use internally for sales people?

    - by Ryan
    I sometimes hear about pain points of customers using our app from sales people, but there really isn't a good way for us to currently keep track of these. I was going to write one myself but figured I would ask first. I was thinking something so simple it would literally just be a small form for adding a new feature, and then it would appear in the list, like stackexchange questions. Then users can upvote them, or even record each time a user complains about something related to the request so we can order them in priority based on real data. Then I can easily go look every few days and see what's going on. That's really it, nothing more complicated than that. Know of anything?

    Read the article

  • I removed Ubuntu from the BIOS menu but it comes back.

    - by jimirings
    For reasons too long to explain, I reluctantly removed Ubuntu from my computer. After completely removing it and deleting the partition that it was installed onto, I discovered that I still had two Ubuntu entries in the boot order in my BIOS menu. I deleted them by following the instructions in this answer: http://askubuntu.com/a/63613/54934 As I was doing it, everything appeared to go smoothly. However, upon reboot one of them came back. What's going on here? How do I delete it permanently? I'll gladly provide any other information that may be needed to diagnose the problem. Thanks.

    Read the article

  • If you want to make money with apps, go BlackBerry

    An interesting study caught my eye. When taken at face value, it provesthat in order to really make money with apps you have to switch to the BlackBerryplatform.A company I never heard of, MPlayit (its URLcuriously opens a Facebook page), is apparently creating an “app discovery app” on Facebookand polled its users about how much they’re willing to spend on apps.Note immediately that this poll represents only the top of the market: the participants are people that areso interested in apps that they’re...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Mass Delete Using Gridview with Checkboxes

    This sample shows how to use a Gridview to delete multiple records all at once, having marked them with a Checkbox, and clicking one button, external to the Gridview, to delete them all As usual, here, we're using the Northwind Database. If you want to try this on your own Northwind database, adhere to this word of caution - "BACK IT UP FIRST". You will be making multiple deletes, so in order to get it back to the original state, you must make a backup and then restore afterwards. Remember too, that you will need to substitute your own Web.Config connectionstring entry for "YourNorthwindString" in the sample code

    Read the article

  • Processing a stream. Must layers be violated?

    - by Lord Tydus
    Theoretical situation: One trillion foobars are stored in a text file (no fancy databases). Each foobar must have some business logic executed on it. A set of 1 trillion will not fit in memory so the data layer cannot return a big set to the business layer. Instead they are to be streamed in 1 foobar at a time, and have business logic execute on 1 foobar at a time. The stream must be closed when finished. In order for the stream to be closed the business layer must close the stream (a data operation detail), thus violating the separation of concerns. Is it possible to incrementally process data without violating layers?

    Read the article

  • Grub-Customizer made Windows 7 disappear from Gurb2

    - by user199055
    I am running dual-boot Ubuntu 12.04/Windows7. I wanted to change the boot order in Grub2 to have Windows on top of the list. I installed Grub-Customizer on Ubuntu 12.04 using the following terminal commands copied from another post: sudo add-apt-repository ppa:adabbas/1stppa sudo apt-get update sudo apt-get install grub-customizer After installing, I launched Grub-Customizer from Dash Home. Windows 7 does not appear in the list. I closed the program and was prompted to save the current settings. Upon rebooting, Windows did not appear in the Grub2 list. I went back into Grub-Customizer and clicked the "Revert" button but still no Windows 7. Does anyone know how I can fix this?

    Read the article

  • looking for advice regarding a free shopping cart solution [on hold]

    - by thirdCharm
    I am building a very small e-commerce site, and I need a free and simple to use/deploy/integrate shopping cart that I can add into my website in order to be able to sell a FEW items. I want the shopping cart to be an add-on in my website, nothing fancy.Ideally when a person clicks on the "Add to cart" button, they will be redirected to the shopping cart, which will then handle different types of payment methods, and everything else you would expect from a fully working shopping cart. I am currently developing my website using the following tools/frameworks: SQL Server 2008 R2 Visual Studio 2010 (ASP.NET 4.5 - C#) HTML5,CSS3, and JS. I am interested in also using PayPal alongside my shopping cart. Help of any kind is appreciated!!

    Read the article

  • What does the 'Burst Rate' stat mean in HDTune?

    - by UpTheCreek
    I recently upgraded my laptop's v slow hard drive to a seagate momentus 7200. Everything is working fine, but I'm a bit confused by these benchmark results: The burst rate is significantly less than the Maximim transfer rate, and not much higher than the normal minimum (if you ignore the spikes). What's going on here? On the HDtune website it defines Burst Rate as: ...the highest speed (in megabytes per second) at which data can be transferred from the drive interface (IDE or SCSI for example) to the operating system. Which begs some questions... e.g. if this is the highest, then how did the bechmarking tool record the 103MB/sec maximum? And if this really is the true maximum, then where is the bottleneck? The laptops SATA interface is on an Intel 82801GBM southbridge controller. When I check in hardware manager, I see that it's driver is iaStor.sys from 2005. Maybe that's the issue? I'll look for a newever version, but any insights would be appreciated. Thanks

    Read the article

  • Run a script with user interaction on log out / shutdown?

    - by lumbric
    I'd like to run a script on shut down, which interrupts the logout process and and pops up a window with zenity. My target is to get autofsck working with lightdm. It seems to work with 12.04 after installing the old *.deb file, if one runs the check script manually. In order to use it, it should run automatically on log out and ask the user if she/he wants to check the disc on shutdown. There is the option session-cleanup-script in the file /etc/lightdm/lightdm.conf which seems to work, if a bash file with full path is used (I can't place the command directly there). But if I press shutdown, there is no time for a user choice. Is there any other option to solve this problem?

    Read the article

  • Using NPM to share resources between UI projects [on hold]

    - by guy mograbi
    I am a UI team leader. My team has a lot of projects using different languages/technologies. In some parts we will rewrite (gradually - @Ampt this is for you) the application in order to enable new fresh technologies in and get old dinosaurs out. I am going to use Node Package Manager to set up an "all powerful" build/dependency manager. Can I use NPM to depend on a private github repository? Can I use NPM to depend on SVN? Will NPM play nice with quickbuild? Since each project might have a slightly different structure (think jetty/maven or play!framework) can I configure NPM to install some dependencies in different folders while still running it from the project's root? How can I, using NPM, get development resources out and build a packaged product? (like a war) Yes/No - is there a reason to use grunt? No discussion, just one liners.

    Read the article

  • How much time do you spend actually developing vs. infrastructure activites?

    - by Can't Tell
    When I'm working I feel like most of the time I'm not doing actual work. For example after making a change to the code in order to test it, I have to first build the project, and start the server(say JBoss). Upon testing, I find that there is another small issue. So I bring down the server, make the changes, build again and start up the server again.The building and bringing the server up/down is not very useful work. Also, the IDE (lets say Eclipse) does things such as updating Maven indexes and building the workspace which take some more time to get things done. Have you come across this kind of situation? Do you have tips on how to overcome/bypass this? Any features on the IDE/build tools that can be helpful? Any architecture/application design/technology that attempts to overcome this?

    Read the article

  • What are the typical methods used to scale up/out email storage servers?

    - by nareshov
    Hi, What I've tried: I have two email storage architectures. Old and new. Old: courier-imapds on several (18+) 1TB-storage servers. If one of them show signs of running out of disk space, we migrate a few email accounts to another server. the servers don't have replicas. no backups either. New: dovecot2 on a single huge server with 16TB (SATA) storage and a few SSDs we store fresh mails on the SSDs and run a doveadm purge to move mails older than a day to the SATA disks there is an identical server which has a max-15min-old rsync backup from the primary server higher-ups/management wanted to pack in as much storage as possible per server in order to minimise the cost of SSDs per server the rsync'ing is done because GlusterFS wasn't replicating well under that high small/random-IO. scaling out was expected to be done with provisioning another pair of such huge servers on facing disk-crunch issues like in the old architecture, manual moving of email accounts would be done. Concerns/doubts: I'm not convinced with the synchronously-replicated filesystem idea works well for heavy random/small-IO. GlusterFS isn't working for us yet, I'm not sure if there's another filesystem out there for this use case. The idea was to keep identical pairs and use DNS round-robin for email delivery and IMAP/POP3 access. And if one the servers went down for whatever reasons (planned/unplanned), we'd move the IP to the other server in the pair. In filesystems like Lustre, I get the advantage of a single namespace whereby I do not have to worry about manually migrating accounts around and updating MAILHOME paths and other metadata/data. Questions: What are the typical methods used to scale up/out with the traditional software (courier-imapd / dovecot)? Do traditional software that store on a locally mounted filesystem pose a roadblock to scale out with minimal "problems"? Does one have to re-write (parts of) these to work with an object-storage of some sort - such as OpenStack object storage?

    Read the article

  • Factory for arrays of objects in python

    - by Vorac
    Ok, the title might be a little misleading. I have a Window class that draws widgets inside itself in the constructor. The widgets are all of the same type. So I pass a list of dictionaries, which contain the parameters for each widget. This works quite nicely, but I am worried that the interface to callers is obfuscated. That is, in order to use Window, one has to study the class, construct a correct list of dictionaries, and then call the constructor with only one parameter - widgets_params. Is this good or bad design? What alternatives does the python syntax provide?

    Read the article

  • Unable to boot from USB boot drive for AMD Athlon 64

    - by Nagarjun
    I am trying to install Ubuntu 13.04 ISO for AMD64 Athlon processor using USB drive (8GB). I did create a bootable USB using the USB installer from Win7 intel 32bit processor (followed steps from here: http://www.ubuntu.com/download/desktop/create-a-usb-stick-on-windows). This USB is unable to boot on my AMD64 bit processor. The bios is able to detect the USB and once I change the boot order to USB and save exit from bios, I get a message saying "DISK BOOT FAILURE, INSERT SYSTEM DISK AND PRESS CONTINUE" and unable to proceed further. However earlier to this I have successfully installed Ubuntu 13.04 on to my laptop which has Intel 32 bit core i5 processor with the same procedure but my desktop which has AMD64 is failing. I also tried to create the USB installer on my desktop which has the AMD64 bit processor and tried booting which also resulted in the same error. (My desktop currently has Win XP) Please help as I have run out of options. thanks, Nagarjun

    Read the article

  • Upgrading to Java 7u65 breaks my Deployment Rule Set for Oracle applications

    - by Don Atreides
    My company uses an older version of an Oracle application that requires Java 6u45. Naturally we want to be secure, so we use a Deployment Rule Set to specify 6u45 for that internal application and let other applications use 7u60. Now that we're ready to upgrade the Java 7 half to 7u67, the Oracle application breaks with "Deployment Rule Set required version 1.6.0_45 not available." Of course it is available, it just can't find it for some reason. As a test, I specified that JavaTester.org should use 6u45 also and it works fine with no issues. But when I try to use the same configuration (7u67 and 6u45) against the Oracle application it fails every time. If I downgrade to 7u60, it works. 7u65 or higher, it breaks. The Oracle application hasn't changed so it must be something different in how 7u65+ is handling Deployment Rule Sets or pathing or something. I'm at a complete loss. ruleset.xml: <?xml version="1.0"?> -<ruleset version="1.0+"> -<rule> <id location="*.mycorp.com"/> <action version="1.6.0_45" permission="run"/> </rule> -<rule> <id location="http://javatester.org"/> <action version="1.6.0_45" permission="run"/> </rule> </ruleset>

    Read the article

  • Adventures in Drupal multisite config with mod_rewrite and clean urls

    - by moexu
    The university where I work is planning to offer Drupal hosting to staff/faculty who want a Drupal site. We've set up Drupal multisite with clean urls and it's mostly working except for some weird redirects. If you have two sites where one is a substring of the other then you'll randomly be redirected to the other site. I tracked the problem to how mod_rewrite does path matching, so with a config file like this: RewriteCond %{REQUEST_URI} ^/drupal RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule ^(.*)$ /drupal/index.php?q=$1 [last,qsappend] RewriteCond %{REQUEST_URI} ^/drupaltest RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_FILENAME} !-d RewriteRule ^(.*)$ /drupaltest/index.php?q=$1 [last,qsappend] /drupaltest will match the /drupal line and all of the links on the /drupaltest page will be rewritten to point to /drupal. If you put the end of string character ($) at the end of each rewrite condition then it will always match on the correct site and the links will always be rewritten correctly. That breaks down as soon as a user logs in though because the query string is appended to the url so just the base url will no longer match. You can also fix the problem by ordering the sites in the config file so that the smallest substring will always be last. I suggested storing all of the sites in a table and then querying, sorting, and rewriting the config file every time a Drupal site is requested so that we could guarantee the order. The system administrator thought that was kludgy and didn't address the root problem. Disabling clean urls should also fix the problem but the users really want them so I'd prefer to keep them if possible. I think we could also fix it by using an .htaccess file in each site to handle the clean url rewriting but that also seems suboptimal since it will generate a higher load on the server and the server is intended to host the majority of the university's external facing web content. Is there some magic I can do with mod_rewrite to get it to work? Would another solution be better? Am I doing something the wrong way to begin with?

    Read the article

  • Renaming hundreds of files at once for proper sorting

    - by Mew
    I have a ton of files, all named stuff like 1.jpg, 2.jpg, 3.jpg, and so on up to 1439.jpg. However, I have a problem with one of my projects and alphabetizing. It will usually go in the order 1.jpg, 10.jpg, 11.jpg and so on. What I need is some way (or a script) to name the files so they are in the format such as 00001.jpg all the way up to 01439.jpg. How would I be able to do this quickly and efficiently?

    Read the article

  • Randomly spawning bitmaps on cnvas

    - by Toystoj
    I need some ideas in order to finish algorithm. I'm randomly placing objects (bitmaps) on canvas without overlapping. Time needed to finish it is my problem. When I need to spawn for example 80% of canvas it takes to long. So i was thinking : I should make some change when the bitmaps take off 50 % of canvas. I want to tell algorithm that it should generate new locations (x,y) where it is free space. My question is : How to render new location (x,y) in place where is free space. In summary: Things I know : object location (x,y) 4 corners (x,y) of object object width, height canvas width, height Any suggestions?

    Read the article

  • LVS / IPVS difference in ActiveConn since upgrading

    - by Hans
    I've recently migrated from an old version of LVS / ldirectord (Ultra Monkey) to a new Debian install with ldirectord. Now the amount of Active Connections is usually higher than the amount of Inactive Connections, it used to be the other way around. Basically on the old load balancer the connections looked something like: -> RemoteAddress:Port Forward Weight ActiveConn InActConn -> 10.84.32.21:0 Masq 1 12 252 -> 10.84.32.22:0 Masq 1 18 368 However since migrating it to the new load balancer it looks more like: -> RemoteAddress:Port Forward Weight ActiveConn InActConn -> 10.84.32.21:0 Masq 1 313 141 -> 10.84.32.22:0 Masq 1 276 183 Old load balancer: Debian 3.1 ipvsadm 1.24 ldirectord 1.2.3 New load balancer: Debian 6.0.5 ipvsadm 1.25 ldirectord 1.0.3 (I guess the versioning system changed) Is it because the old load balancer was running a kernel from 2005, and ldirectord from 2004, and things have simply changed in the past 7 - 8 years? Did I miss some sysctl settings that I should be enforcing for it to behave in the same way? Everything appears to be working fine but can anyone see an issue with this behaviour? Thanks in advance! Additional info: I'm using LVS in masquerading mode, the real servers have the load balancer as their gateway. The real servers are running Apache, which hasn't changed during the upgrade. The boxes themselves show roughly the same amount of Inactive Connections shown in ipvsadm.

    Read the article

  • Programming Entity Framework, 2nd Edition (EF4) Table of Contents

    We are closing in on finalizing the 2nd edition of Programming Entity Framework! Although the rough draft chapters are already available through Safari’s Rough Cuts program (here: http://oreilly.com/catalog/9780596807252) I have been editing and reshaping the content since those chapters were published. You can get the final print edition (August 15th or perhaps a bit earlier) at O’Reilly or pre-order it here on Amazon.com (here) (and elsewhere of course!) I believe that the book will...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Validation and Verification explanation (Boehm) - I cannot understand its point

    - by user970696
    Hopefully my last thread about V&V as I found the B.Boehm is text which I just do not understand well (likely my technical English is not that good). http://csse.usc.edu/csse/TECHRPTS/1979/usccse79-501/usccse79-501.pdf Basically he says that verification is about checking that products derived from requirements baseline must correspond to it and that deviation leads only to changes in these derived products (design, code). But he says it begins with design and ends with acceptance tests (you can check the V model inside). The thing is, I have accepted ISO12207 in terms of all testing is validation, yet it does not make any sense here. In order to be sure the product complies with requirements (acceptance test) I need to test it. Also it says that validation problems means that requirements are bad and needs to be changed - which does not happen with testing that testers do, who just checks correspondence with requirements.

    Read the article

  • Fatal Error when booting

    - by Denja
    I'm running Ubuntu 10.10 with ATI proprietary FGLRX graphic drivers When I boot, I can see this very quick message appearing; I don't have time to read it: **Fatal Error ..................................**(&@something i cant read) I searched through the log file in /var/log/ in order to find what is wrong and I did find something in the /var/log/Xorg.1.log: 21:31:08 [ 15.734] (--) using VT number 1 [ 204.647] Fatal server error: [ 204.647] xf86OpenConsole: VT_WAITACTIVE failed: Interrupted system call [ 204.647] [ 204.647] Please consult the The X.Org Foundation support at http://wiki.x.org for help. [ 204.647] Please also check the log file at "/var/log/Xorg.1.log" for additional information. [ 204.647] But this is already Xorg.1.log. And there is a Xorg.0.log also & Xorg.0.log.old but it doesnt have any error in it. My system seems to work properly and it seems its not affected by this But how do I correct this message? Any suggestions?

    Read the article

< Previous Page | 484 485 486 487 488 489 490 491 492 493 494 495  | Next Page >