Search Results

Search found 64952 results on 2599 pages for 'windows live sync'.

Page 1856/2599 | < Previous Page | 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863  | Next Page >

  • What gaming keyboard(s) will work with Ubuntu?

    - by belacqua
    I've been looking at gaming keyboards to use on Ubuntu system. Microsoft has a few popular ones (e.g., Sidewinder X4, X6), but the programmable function keys appear to be unusable without the Windows software. (Though here's a post from someone who has a more recent project that uses usbmon and xdotool to add functions to some keys.) Another choice in my budget is the Cyborg V.05. It seems about right for my needs, but I would be depressed having a bunch of useless, nonprogrammable keys on it. Logitech has some models (e.g., the Logitech G110), though again I expect that the extensive macro capabilities (which I don't need) would be lost under Linux. There's a project called g15tools which has some code to work with older Logitech gaming models, but I don't know what the current status is. Last entry there was in March 2010. There are also a number of very old posts around the internet with regard to the Logitech G11 and G15. Compatibility with the current keyboards, Ubuntu version, and Linux kernel are suspect. I'm in the U.S., and so it appears that few of the Roccat keyboards are available, and they're over-priced. Support might be OK for these, though -- there's a short Phoronix article about Roccat improving their Linux support, and there's also a project and webpage for "Using Roccat Hardware with Linux". Honestly, the only feature I have to have is good backlighting for the keys, and if it's not wired (which is fine), the wireless capability should function. I could probably live with dead function keys, as long as they weren't in places that would interfere with things like Unity/compiz shortcuts. Any experience or suggestions? I've not seen much to inspire confidence with programmable/macro keys. There is a thread (with no solutions) on the Sidewinder X4 on ubuntuforums here. I'm also considering the Logitech Illuminated Keyboard as a possibility, even though it's not specifically a gaming keyboard. It is backlit, and it's supposed to be a nice keyboard.

    Read the article

  • Fusion Applications: Get Trained to Implement

    - by mseika
    0 0 1 214 1349 Technische Universität München 11 3 1560 14.0 96 800x600 Normal 0 21 false false false DE JA X-NONE /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Normale Tabelle"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:10.0pt; font-family:"Calibri","serif";} Becoming an Oracle Fusion Applications Implementation Expert gives you the opportunity to gain differentiation and a competitive advantage. Oracle University can help you to get trained and then specialized. View the current course schedule for Romania and Turkey. You can choose a delivery format to suit your learning style and situation: Classroom Training in Bucharest, Romania or Istanbul, Turkey or the travel-free options: Live Virtual Class or Training On Demand. SPECIAL OFFER Standard prices for all Oracle Fusion Applications Training On Demand courses currently have a reduced price. This offer is valid until 31 August 2013. Your OPN discount will be then applied to this reduced price to make this learning form even more attractive.

    Read the article

  • AngularJS dealing with large data sets (Strategy)

    - by Brian
    I am working on developing a personal temperature logging viewer based on my rasppi curl'ing data into my web server's api. Temperatures are taken every 2 seconds and I can have several temperature sensors posting data. Needless to say I will have a lot of data to handle even within the scope of an hour. I have implemented a very simple paging api from the server so the server doesn't timeout and is currently only returning data in 1000 units per call, then paging through the data. I had the idea to intially show say the last 20 minutes of data from a sensor (or all sensors depending on user choices), then allowing the user to select other timeframes from which to show data. The issue comes in when you want to view all sensors or an extended time period (say 24 hours). Is there a best practice of handling this large amount of data? Would it be useful to load those first 20 minutes into the live view and then cache into local storage something like the last 24 hours? I haven't been able to find a decent idea of this in use yet even though there are a lot of ways to take this problem. I am just looking for some suggestions as to what might provide a good balance between good performance and not caching the entire data set on the client side (as beyond a week of data this might not be feasible).

    Read the article

  • From release to business

    - by geneotech
    So let's say that I've finished programming a simple, indie MMO game similiar to Tibia. I've got a stable server application that is ready to launch, i've got a tested bug-free working client application that is ready to play and the game's official website (ready to host) with payment system and client that is ready to download for free. Let's say none of them break copyright laws, and no matter how impossible it sounds, let's for now say it's true. My game divides accounts into two groups - free and premium. If someone gets premium, he's granted access to all possible game features, that of course, need server authorisation to work properly. Let's say that the "premium account" can be bought on the website for a fixed money/month. Free accounts mean that everyone can actually play, but without paying, you get limited access. This is what the mentioned payment system will be for. Well, I'm completely novice to these business entities issues, so in short: what, in terms of law, are steps from here to the state where my game earns money in a fully legal way ? Also, is there for example, something like verification if game gives the user what it actually offers when paying on its website ? I live in Europe, if it changes something.

    Read the article

  • Is full partition encryption the only sure way to make Ubuntu safe from external access?

    - by fred.bear
    (By "external access", I mean eg. via a Live CD, or another OS on the same dual-boot machine) A friend wants to try Ubuntu. He's fed up with Vista grinding to a crawl (the kids? :), so he likes the "potential" security offered by Ubuntu, but because the computer will be multi-booting Ubuntu (primary) and 2 Vistas (one for him, if he ever needs it again, and the other one for the kids to screw up (again). However, he is concerned about any non-Ubuntu access to the Ubuntu partitions (and also to his Vista partition)... I believe TrueCrypt will do the job for his Vista, but I'd like to know what the best encryption system for Ubuntu is... If TrueCrypt works for Ubuntu, it may be the best option for him, as it would be the same look and feel for both. Ubuntu will be installed with 3 partitions; 1) root 2) home 3) swap.. Will Ubuntu's boot loader clash with TrueCrypt's encrypted partition? PS.. Is encryption a suitable solution?

    Read the article

  • Best setup/workflow for distributed team to integrated DSVC with fragmented huge .NET site?

    - by lazfish
    So we have a team with 2 developers one manager. The dev server sits in a home office and the live server sits in a rack somewhere handled by the larger part of my company. We have freedom to do as we please but I want to incorporate Kiln DSVC and FogBugz for us with some standard procedures to make sense of our decisions/designs/goals. Our main product is web-based training through our .NET site with many videos etc, and we also do mobile apps for multiple platforms. Our code-base is a 15 yr old fragmented mess. The approach has been rogue .asp/.aspx pages with some class management implemented in the last 6 years. We still mix our html/vb/js all on the same file when we add a feature/page to our site. We do not separate the business logic from the rest of the code. Wiring anything up in VS for Intelli-sense or testing or any other benefit is more frustrating than it is worth, because of having to manually rejigger everything back to one file. How do other teams approach this? I noticed when I did wire everything up for VS it wants to make a class for all functions. Do people normally compile DLLs for page-specific functions that won't be reusable? What approaches make sense for getting our practices under control while still being able to fix old anti-patterns and outdated code and still moving towards a logical structure for future devs to build on?

    Read the article

  • Erfolgreich mit Referenzen

    - by A&C Redaktion
    Erfolgreich umgesetzte Projekte sind die beste Werbung! So wie zum Beispiel der Anwenderbericht mit der DAW-Gruppe, Hersteller der bekannten Marke "Alpinaweiß" sowie vieler anderer Farben, und Oracle Partner Freudenberg IT. Lesen Sie, wie DAW mithilfe der Oracle Database Enterprise Edition 11g und Oracle Advanced Compression die Speicherleistung um 40% erhöhen und bis zu 20% mehr Datenbank-Performance erzielen konnte.  Sie möchten auch einen Anwenderbericht zu einem abgeschlossenen Kundenprojekte haben? So gehts: Schicken Sie einfach ein paar Eckdaten zum Projekt, vorausgesetzt Ihr Endkunde ist einverstanden, an Frau Marion Aschenbrenner, alles weitere übernimmt Oracle - einschließlich der Finanzierung! Folgende Details sollten Sie uns nennen: Kundenname & Ansprechpartner (mit Telefon und Email-Adresse) Eingesetzte Oracle Produkte Kurze Beschreibung des Projektes (3-4 Sätze mit dem Hauptnutzen der Oracle Lösung) Ansprechpartner bei Ihnen im Hause für das Kundenprojekt und alles, was noch wichtig zu wissen ist (Go-Live-Datum, Terminvorschläge, etc.) Nach einer erfolgreichen Prüfung Ihrer Nominierung werden Sie und der Endkunde per Telefon interviewt und ein ausführlicher Anwenderbericht erstellt. Diesen bekommen Sie und ihr Kunde natürlich zur Freigabe vorgelegt. Veröffentlicht werden die Referenzberichte unter anderem hier im Blog und in anderen sozialen Medien.

    Read the article

  • Erfolgreich mit Referenzen

    - by A&C Redaktion
    Erfolgreich umgesetzte Projekte sind die beste Werbung! So wie zum Beispiel der Anwenderbericht mit der DAW-Gruppe, Hersteller der bekannten Marke "Alpinaweiß" sowie vieler anderer Farben, und Oracle Partner Freudenberg IT. Lesen Sie, wie DAW mithilfe der Oracle Database Enterprise Edition 11g und Oracle Advanced Compression die Speicherleistung um 40% erhöhen und bis zu 20% mehr Datenbank-Performance erzielen konnte.  Sie möchten auch einen Anwenderbericht zu einem abgeschlossenen Kundenprojekte haben? So gehts: Schicken Sie einfach ein paar Eckdaten zum Projekt, vorausgesetzt Ihr Endkunde ist einverstanden, an Frau Marion Aschenbrenner, alles weitere übernimmt Oracle - einschließlich der Finanzierung! Folgende Details sollten Sie uns nennen: Kundenname & Ansprechpartner (mit Telefon und Email-Adresse) Eingesetzte Oracle Produkte Kurze Beschreibung des Projektes (3-4 Sätze mit dem Hauptnutzen der Oracle Lösung) Ansprechpartner bei Ihnen im Hause für das Kundenprojekt und alles, was noch wichtig zu wissen ist (Go-Live-Datum, Terminvorschläge, etc.) Nach einer erfolgreichen Prüfung Ihrer Nominierung werden Sie und der Endkunde per Telefon interviewt und ein ausführlicher Anwenderbericht erstellt. Diesen bekommen Sie und ihr Kunde natürlich zur Freigabe vorgelegt. Veröffentlicht werden die Referenzberichte unter anderem hier im Blog und in anderen sozialen Medien.

    Read the article

  • Fusion Applications: Get Trained to Implement

    - by mseika
    0 0 1 165 1041 Technische Universität München 8 2 1204 14.0 96 800x600 Normal 0 21 false false false DE JA X-NONE /* Style Definitions */ table.MsoNormalTable {mso-style-name:"Normale Tabelle"; mso-tstyle-rowband-size:0; mso-tstyle-colband-size:0; mso-style-noshow:yes; mso-style-priority:99; mso-style-parent:""; mso-padding-alt:0cm 5.4pt 0cm 5.4pt; mso-para-margin:0cm; mso-para-margin-bottom:.0001pt; mso-pagination:widow-orphan; font-size:10.0pt; font-family:"Calibri","serif";} Becoming an Oracle Fusion Applications Implementation Expert gives you the opportunity to gain differentiation and a competitive advantage. Oracle University can help you to get trained and then specialized. View the current course schedule for France. You can choose a delivery format to suit your learning style and situation: Classroom Training in Paris Colombes or the travel-free options: Live Virtual Class or Training On Demand. SPECIAL OFFER Standard prices for all Oracle Fusion Applications Training On Demand courses currently have a reduced price. This offer is valid until 31 August 2013. Your OPN discount will be then applied to this reduced price to make this learning form even more attractive.

    Read the article

  • It Ain't Over 'Til It's Over

    - by Oracle OpenWorld Blog Team
    Oracle OpenWorld 2012 is behind us. Well, for San Francisco, anyhow. The team is already working on the Latin America event which takes place in December in Sao Paulo, and an OpenWorld in Asia for 2013 as well. And of course they're already working on the next San Francisco OpenWorld for 2013. So what happens after the conference is over? People pack up demo and network gear and ship it out to wherever it's going next; take down and recycle signage; strike the keynote set, the exhibition and demo halls, the street tents, and anything else that was constructed just for the conference. There's a lot of post-conference analyis going on too. Oracle and partner marketing teams are looking at and following up on the leads they got from booth, demo, and lounge traffic. The events team is evaluating the session and conference surveys you filled out if you attended -- looking to identify the best speakers, what worked and didn't work, how you liked the venues, the food, the entertainment, the presentations. From all of that information will come recommendations for next year on what to keep doing, what to do better, and what not to do at all. The goal for each year's conference is to be better than last year's. If you attended and haven't filled out the surveys yet, you have until October 19 for them to be counted, and for you to be entered into a daily sweepstakes. Click here for more information. Posts to this blog will slow down for a while, but we'll post news about Oracle OpenWorld in San Francisco and around the world when we have it. Any suggestions about future blog topics are welcome. Oh - I forgot to mention that you can sign up to be notified when registration for Oracle OpenWorld 2013 goes live. If you register at that time you'll get the best discount available on attending next year. So sign up, and stay tuned.

    Read the article

  • A Brand-new Automated Testing Tool is the Result of Telerik and ArtOfTest Merger

    Im sure youve already heard the great news about Telerik expansion and the new Telerik Automated Testing Tools division. I am excited to share what we worked on and produced for the last couple of months. New Release The latest Telerik release that went live this week added a completely new tool to Teleriks automated testing product line. The new QA Edition is tailored for QA Professionals. The QA Edition is a standalone tool that allows QAs to freely create, execute and maintain their tests without having to install Visual Studio. If you are a developer and you want something much faster and lightweight than VS, then the Standalone tool is worth trying. New IDE The QA Edition is a WPF application with interface built on top of the latest and greatest RadControls for WPF. This allowed us to configure and build intuitive and easy-to-use UI. Additionally, the rich ...Did you know that DotNetSlackers also publishes .net articles written by top known .net Authors? We already have over 80 articles in several categories including Silverlight. Take a look: here.

    Read the article

  • Resized Ubuntu 14.04 partition will not boot

    - by user292577
    First, a little background info: I intended to install Ubuntu alongside OS X, but I accidentally erased my entire hard drive and was left with Ubuntu alone. Yesterday, I finally tried to create a desperate partition and OS X. I successfully shrank my Ubuntu partition and created ~40 GB of free space. I used gparted to do this. At first I tried to use Internet Recovery to install OS X on the unallocated space I had created, but discovered I couldn't. I went back to gparted and created a FAT partition with that space. I'm pretty sure I actually did this using my Ubuntu partition on my hard drive (the one I had just shrunken), rather than the USB live boot I had used to shrink it. Therefore, at this point, I think my Ubuntu partition was still functioning properly. I went back to internet recovery and used disk utility to turn the FAT partition into a Mac OS Extended (Journaled) [HFS/HFS+] partition. I believe this is probably where the problem occurred. I successfully reinstalled OS X and found it to be fully functioning. However, when I tried to boot back into the Ubuntu partition all I got was a black screen with a little white cursor (it looks similar to a terminal screen/the Mac DFU mode, but I can't type anything into it. Is there anything I can do to repair my Ubuntu partition? If not, can I at least recover the data from it? Thank you for any help.

    Read the article

  • Seek first to understand, then to be understood

    - by BuckWoody
    One of the most important (and most difficult) lessons for a technical professional to learn is to not jump to the solution. Perhaps you’ve done this, or had it happen to you. As the person you’re “listening” to is speaking, your mind is performing a B-Tree lookup on possible solutions, and when the final node of the B-Tree in your mind is reached, you blurt out the “only” solution there is to the problem, whether they are done or not. There are two issues here – both of them fatal if you don’t factor them in. First, your B-Tree may not be complete, or correct. That of course leads to an incorrect response, which blows your credibility. People will not trust you if this happens often. The second danger is that the person may modify their entire problem with a single word or phrase. I once had a client explain a detailed problem to me – and I just KNEW the answer. Then they said at the end “well, that’s what it used to do, anyway. Now it doesn’t do that anymore.” Which of course negated my entire solution – happily I had kept my mouth shut until they finished. So practice listening, rather than waiting for your turn to speak. Let the person finish, let them get the concept out, give them your full attention. They’ll appreciate the courtesy, you’ll look more intelligent, and you both may find the right answer to the problem. Share this post: email it! | bookmark it! | digg it! | reddit! | kick it! | live it!

    Read the article

  • Crackling sound from right laptop speaker

    - by user1880405
    This problem lasts for several months already (first on Ubuntu 13.10, not on 12.04. I get very loud cracking/popping sound from my right Asus K56C's speaker, I searched everything but could not get rid of it. Several facts: There is no problem on Windows 8. It has nothing to do with applications running because it appears even before login screen of Ubuntu. Also same problem if I boot from Live USB. Muting sound will remove noise, but lowering volume has no effect. Inserting any headphones, removes the noise. If I disconnect power cable while there is noise, noise will always disappear, but only if there is no music playing. If I start playing music, noise again appear even with power cable disconnected. Sometimes that noise disappears for 1-4 weeks, and then again appears for no reason and lasts from several days to weeks. That noise is always the same, and I tried adding tsched=0 to /etc/pulse/default.pa. Also tried this PositionReporting fix, with no effect. I also tried disconnecting all the cables and removing all electronic devices around laptop, but it has no effect. I also tried removing Pulse Audio, didn't change anything. Would be great if someone has some real solution for this problem.

    Read the article

  • Need some critique on .NET/WCF SOA architecture plan

    - by user998101
    I am working on a refactoring of some services and would appreciate some critique on my general approach. I am working with three back-end data systems and need to expose an authenticated front-end API over http binding, JSON, and REST for internal apps as well as 3rd party integration. I've got a rough idea below that's a hybrid of what I have and where I intend to wind up. I intend to build guidance extensions to support this architecture so that devs can build this out quickly. Here's the current idea for our structure: Front-end WCF routing service (spread across multiple IIS servers via hardware load balancer) Load balancing of services behind routing is handled within routing service, probably round-robin One of the services will be a token Multiple bindings per-service exposed to address JSON, REST, and whatever else comes up later All in/out is handled via POCO DTOs Use unity to scan for what services are available and expose them The front-end services behind the routing service do nothing more than expose the API and do conversion of DTO<-Entity Unity inject service implementation to allow mocking automapper for DTO/Entity conversion Invoke WF services where response required immediately Queue to ESB for async WF -- ESB will invoke WF later Business logic WF layer Expose same api as front-end services Implement business logic Wrap transaction context where needed Call out to composite/atomic services Composite/Atomic Services Exposed as WCF One service per back-end system Standard atomic CRUD operations plus composite operations Supports transaction context The questions I have are: Are the separation of concerns outlined above beneficial? Current thought is each layer below is its own project, except the backend stuff, where each system gets one project. The project has a servicehost and all the services are under a services folder. Interfaces live in a separate project at each layer. DTO and Entities are in two separate projects under a shared folder. I am currently planning to build dedicated services for shared functionality such as logging and overload things like tracelistener to call those services. Is this a valid approach? Any other suggestions/comments?

    Read the article

  • From the Coalface - 3 - Work as hard as you can to be as lazy as you can!

    - by TATWORTH
    The saga of the Change Log A recent conversation reminded me of the need for change logs within a database, to record when various change scripts were run. Creating such the required table is simple. A typical table for this consists of: Id - identity Integer primary key ChangeFileName - NVARCHAR(128) to hold the name of the file run. DateAdded - DateTime non-null with default value of getutcdate() Purpose - NVARCHAR(128) Rerunnable - Bit non-null default 0. By good design of the table only two data values normally need to be supplied. Two stored procedures, one for inserting data and one to list in reverse sequence the log complete the database essentials. The complete implementation can be found in the CommonData solution at http://CommonData.CodePlex.Com By including a call the add Change Log stored procedure, each script can log its name and purpose for posterity. The scripts that were applied to say the UAT system and their sequence of application can be readily identified for running on the Live system. Formatting XML XML is often produced as one continous string with no embedded CR/LF. To get it into human readable form, open it in visual studio, swap to another tab and back and click the format document button. The XML will then be nicely formatted!

    Read the article

  • 2 year cis degree and in school for computer science what can I do?

    - by chame1eon
    Hi I am 29 and have a recent 2 cis year degree from a community college , an A+ certification , and meager experience with web stuff ( Java , Javascript , php ) while in my 1 year help desk internship. In all the programming classes I was able to blow through the homework easily even while other students were panicking and dropping. I think I have managed to avoid the most atrocious noob/self taught mistakes ( spaghetti code etc) by just doing research before starting something and trying to keep good design in mind. Even so I'd have to make heavy use of references to crawl through even simple projects that would result in fully finished useful applications. I need a job now and I am tired of the slow pace of the classes and would love to get any kind of practical experience I could. The problem is that I am not sure what I should be trying to do. I have a very strong preference for application programming or at least anything light on design and preferably pretty low level. If I can't do that then anything technology related , for example help desk would be better than nothing. I live near Raleigh NC. Am I qualified for anything that could contribute to coding (C++ or Java ) experience or even web development though I don't really like it. Would web development experience help. If not is there anything I could read or do that could help? Is the help desk my only choice? If it is, are there any relatively quick certifications or anything similar that would help while I am waiting? Sorry about the long multi-part question. Thanks for reading.

    Read the article

  • Do you have a data roadmap?

    - by BuckWoody
    I often visit companies where they asked me “What is SQL Server’s Roadmap?” What they mean is that they want to know where Microsoft is going with our database products. I explain that we’re expanding not only the capacities in SQL Server but the capabilities – we’re trying to make an “information platform”, rather than just a data store. But it’s interesting when I ask the same question back. “What is your data roadmap?” Most folks are surprised by the question, thinking only about storage and archival. To them, data is data. Ah, not so. Your data is one of the most valuable, if not the most valuable asset in your organization. And you should be thinking about how you’ll acquire it, how it will be distributed, how you’ll archive it (which includes more than just backing it up) and most importantly, how you’ll leverage it. Because it’s only when data becomes information that it is truly useful. to be sure, the folks on the web that collect lots of data have a strategy for it – do you? Share this post: email it! | bookmark it! | digg it! | reddit! | kick it! | live it!

    Read the article

  • Oracle Linux Partner Pavilion Spotlight III

    - by Ted Davis
    Three days until Oracle OpenWorld 2012 begins. The anticipation and excitement are building. In today's spotlight we are presenting an additional three partners exhibiting in the Oracle Linux Partner Pavilion at Oracle OpenWorld ( Booth #1033). Fujitsu will showcase a Gold tower system representing the one-millionth PRIMERGY server shipped, highlighting Fujitsu’s position as the #4 server vendor worldwide. Fujitsu’s broad range of server platforms is reshaping the data center with virtualization and cloud services, including those based on Oracle Linux and Oracle VM. BeyondTrust, the leader in providing context aware security intelligence, will be showcasing its threat management and policy enablement solutions for addressing IT security risks and simplifying compliance. BeyondTrust will discuss how to reduce security risks, close security gaps and improve visibility across your server and database infrastructure. Please stop by to see live demonstrations of BeyondTrust’s award winning vulnerability management and privilege identity management solutions supported on Oracle Linux. Virtualized infrastructure with Oracle VM and NetApp storage and data management solutions provides an integrated and seamless end user experience. Designed for maximum efficiency to allow for native NetApp deduplication and backup/recovery/cloning of VM’s or templates. Whether you are provisioning one or multiple server pools or dynamically re-provisioning storage for your virtual machines to meet business demands, with Oracle and NetApp, you have one single point-and-click console to rapidly and easily deploy a virtualized agile data infrastructure in minutes. So there you have it!  The third install of our Partner Spolight. Check out Part I and Part II of our Partner Spotlights from previous days if you've missed them. Remember to visit the Oracle Linux team at Oracle OpenWorld.

    Read the article

  • Top down or bottom up approach?

    - by george_zakharov
    this is the closest I think I can get to define my problem. I'm an interface designer and now I'm working with a team of programmers to develop a new CMS for a heavy media site. I'm sorry if it's a very, very dumb question to ask here, but I really need some help. As I've started developing a specification list for a prototype it turned out a very big one. I do realize now that the client-side will be JS heavy, with lots of DnD and other "cool designer stuff". The CMS will even include a small project management system for its users, nothing big like Basecamp, but with a live feed of comments etc. The problem is the the team has now separated. Someone is proposing the existing backend solution used in other CMS, someone is proposing to rewrite everything from scratch. The point to keep the code is that it is faster, the point to rewrite is to make it better for the proposed design (include Node.js and other stuff I don't actually get). The question is — can the UI specs influence back-end? The guys that propose to use the existing solution did everything with the Yii framework (as far as I know), and they say that everything on server is not affected by this "interface coolness". Others say that it does, that even autosave can't work without server load. Please, if this is really incomprehensible, again, I'm sorry, and I'll happy to clarify it after your questions. Thanks in advance

    Read the article

  • Using Subdomains for Newly Regional Company

    - by Taylord22
    The company I work for is expanding their business to new territories. I've got a lot of stabilization to do in the region/state where we're one of the most well known companies of our kind. Currently, we have 3 distinct product lines which are currently distinguished by 3 separate URLS. This is affecting the user flow of our site, so we'd like to clean it up before launching our products into the various regions. The business has decided to grow into 5 new states (one state consisting of one county only) — none of which will feature all 3 products. Our homebase state is the only one that will have all 3 products this year. My initial thought was to use subdomains to separate out the regions, that way we could use a canonical tag to stabilize the root domain (which would feature home state content, and support content for all regions), and remove us from potential duplicate content penalization. Our product content will be nearly identical across the regions for the first year. I second guessed myself by thinking that it was perhaps better to use a "[product].root/region" URL instead. And I'm currently stuck by wondering if it was not better to build out subdomains for products and regions...using one modifier or the other as a funnel/branding page into the other. For instance, user lands on "region.root.com" and sees exactly what products we offer in that region. Basically, a tailored landing page. Meanwhile the bulk of the product content would actually live under "product.root.com/region/page". My head is spinning. And while searching for similar questions I also bumped into reference of another tag meant to be used in some similar cases to mine. I feel like there's a lot of risks involved in this subdomain strategy, but I also can't help but see the benefits in the user flow.

    Read the article

  • When you’re on a high, start something big

    - by BuckWoody
    Most days are pretty average – we have some highs, some lows, and just regular old work to do. But some days the sun is shining, your co-workers are especially nice, and everything just falls into place. You really *enjoy* what you do. Don’t let that moment pass. All of us have “big” projects that we need to tackle. Things that are going to take a long time, and a lot of money. Those kinds of data projects take a LOT of planning, and many times we put that off just to get to the day’s work. I’ve found that the “high” moments are the perfect time to take on these big projects. I’m more focused, and more importantly, more positive. And as the quote goes, “whether you think you can or you think you can’t, you’re probably right.” You’ll find a way to make it happen if you’re in a positive mood. Now – having those “great days” is actually something you can influence, but I’ll save that topic for a future post. I have a project to work on. :) Share this post: email it! | bookmark it! | digg it! | reddit! | kick it! | live it!

    Read the article

  • Poor performance after reinstalling to a USB drive

    - by anonymous
    I am currently running Ubuntu 11.10 off of a SanDisk 16GB USB. I installed it using a Live USB with the following partition configuration: 6GB Primary /dos FAT32 5GB Logical / ext4 5GB Logical /home ext4 I don't have a hard disk, and don't see myself getting one anytime soon. I rely solely on this 16GB, and two other 4GB USBs, one of which I used as the LiveUSB. I bring the USBs around, and even use the install at work. I previously used an install that used a swap file. It functioned fine for the most part, save for a few slow moments, but I came across this post, and it got me thinking about my USB's life, so I reinstalled with the current config. My problem now is that it is slower. Applications like Firefox would hang more often. In my previous setup (the automatically partitioned setup), Firefox would start hanging if I was running an unzip or install task on the same partition as /. Now however, it would hang if I had another window open i.e. the system settings window. My guess is that it may have something to do with the swap file or the install being on a Logical partition rather than a Primary partition, but I don't know. Any insight as to why it has slowed down?

    Read the article

  • Don't speak to the DBA while he/she is doing the job?!

    - by Uri Dimant
    Our DBA was very busy today, he helped out our developers to write efficient code,explained to the programmer how to launch the old DTS package on SQL Server 2005 and etc. You know how it is, someone comes in and asking the question and you answer the question at the same time you are busy to do something. So  today answering one of such questions he deleted by mistake very ctitical database. Fortunately, we had zero data loss,thanks to our backup/recovery strategy. I think that is acceptable to say that you are busy right now ,please ask the question later on, what do you think? How do you respond if someone asks you a queston while you are doing something on the server/database and etc.?   Share this post: email it! | bookmark it! | digg it! | reddit! | kick it! | live it!

    Read the article

  • trying to upgrade memory

    - by user214876
    I've been using Ubuntu on my laptop for awhile now. Not quite used to it yet. I've got a Acer Aspire with an orig 4 gig mem/500 gig HDD. Running 12.04 presently 32 bit sys. I have the 13.04 upgrade disc and want to upgrade my memory to 8 Gig. Everytime I install the 8 gig memory, the system won't boot to either version. I downloaded the 64 bit version of both versions of Ubuntu but no results yet. Can anyone offer a suggestion here? I'm kinda lost. Additional Information: The memory was purchased through Acer/Kingston. Recommended for this computer. I watched the video on installing it, so I doubt it's installed wrong. (There's only one way of putting it in). I swapped Op Sys, from Ubuntu 12.04 to 13.04 to 13.10 and now to Xubuntu 13.10 64 bit version. I'm still not having any luck with this upgrade. Would it be necessary to upgrade the CPU? It's just a thought, I don't know what else could keep me from utilizing the new memory. Additional Information: Called Kingston this afternoon, they are sending replacement lower density memory modules 2/4 gig - 8 gig. Tech service says I need to upgrade BIOS to utilize new memory install VIA Dos since it is no longer a windows system. I'm not sure how to go about that but it's a learning process I can live with. Thank you all for your help/support. I realize this isn't a Ubuntu problem but each new user of this op sys, seems to share simular problems and maybe someone can use this info to their advantage.

    Read the article

< Previous Page | 1852 1853 1854 1855 1856 1857 1858 1859 1860 1861 1862 1863  | Next Page >