Few weeks ago I went to attend a WeAreDevelopers conference in Vienna. The venue is just 10 minutes walk time from the place where I live.
I will go through all of the talks I watched with a very brief message before giving my thoughts on the organization of the conference.
Opening. For such a small country as Austria a conference hosting 3500+ developers is a big deal. Chancellor of Austria (most powerful position) took the keyword and, to my disappointment, for the most part spoke in German. I think this would be fine if only this conference wasn’t advertised as an international conference. In any case it is admirable that importance of IT industry is acknowledged by the Austrian government.
Build a World We All Want to Live. This talk was a lot about the future and exponential growth. I’m afraid there not much of takeaway except of a sense of a motivational inspiration.
Challenges of Autonomous Driving. I first heard about Rimac concept car on TopGear TV Show. It was very interesting to hear about the Croatia’s concept car from engineering perspective. What I learned is that there are a lot more of a challenge that has to be addressed than it is advertised by the car makers such as Tesla, Honda, etc.
IoT & Advanced Analytics – Real World Challenges for Developers. Austrian railway company explaining how they use all kinds of detectors on their trains and how their data is analysed. This does not get anywhere close to using IA, though still interesting.
One ID to Rule Them All. This was supposed to be a presentation on identification methods (think passports, ID cards) but somehow the presenter talked about solar panels, exponential growth of technology, and near free energy. Not sure how these two play together, but the title was definitely misleading.
How Different Open Hardware is to Open Software. Interesting talk, but again misleading title. Speaker presented his robotic arm project and how it can be used by people without an arm. Takeaway: these days you can download some code, buy Arduino and 3D-print yourself an arm within hours.
The Early Days of Id Software: Programming Principles. I found zillion of youtube videos with the same title where Romero gives the same talk on different conferences. Best to watch would be this another video here. He goes through the history of Id-software in the talk. Takeaway: John Romero’s Principles for Programmers
Getting Computers to Understand Us. This was a presentation on NLP and AI that started with punchcards.
How to Design Human Centered Chatbots? Takeaway: I didn’t learn how to design chatbots but I understood that chatboat euphoria is on a downtrend.
The Future of Online Money: Creating Secure Payments Globally. This was about PayPal money transactions and security. Takeway: eliminating middleman helps in improving processes.
Less Process, More Guidance. Takeway: The Atlassian Team Playbook.
Extreme Continuous Integration. Automic company presented how they do hundreds of builds of different components for multiple platforms continuously.
Continuous Delivery Journey @ Wirecard. This was more of continuation of CI and CD topic.
Monorepos in the wild. A story of going to one repository. Some pros and cons where presented with a message that monorepo != monoapp.
Rebuilding an Aircraft on the Fly. Yet another story, now about fixing CSS by Trivago company. In my opinion this was just a common sense story.
JS @ Uber. About programming languages & architecture approaches at Uber. Things like NodeJs, Go, Python, and micro-services were mentioned in addition to JS.
The Artist and the Machine. A lightning talk about nice spiral graphs. Takeaway: I bought my daughter a spirograph toy.
Working Backwards from the Customer. Amazon explained how they start developing with a press release. If they don’t like what is written it might not make sense to start doing it. If they like what is written and it sounds cool they add more details, create more technical translation and finally this is converted into development. I believe this was my biggest takeaway from the conference in general. In software projects it happens very often that customer does not get what they want. Starting with customer and being obsessed with the customer is probably something that makes Amazon stand out.
Model-Minded Development. A presenter from Google talked about importance of having a good model between computer and the real world.
Customizing Railways to Individuality. National Austrian railroad company talked about their challenges. They are in business for 178 years and can be considered dinosaurs when it comes to software.
Scaling Open Source Communities. Tips and tricks of handling an open source project were given. A lot of insight into OS software lifecycle.
Angular, Google’s Popular Application Framework. A relatively simple demo of Angular framework was given. Takeaway: Angular will last longer than any other JS framework. Let’s see.
PHP in 2017. This was somewhat hilarious talk by the inventor of PHP. He talked about lots of performance improvements that come with PHP7 and how that can “save the planet”. Takeaway: This blog is self-hosted word-press. I upgraded the PHP version to 7 making my contribution to more sustainable future.
Developers Are Writing the Script for the Future. Definitely the highlight of the conference was a presentation by the creator of StackOverflow Joel Spolsky. This was inspirational talk and a great way to close the conference.
Many people have complained about significant overbooking for the conference. It was very unpleasant to see organizers bragging about 3,8K attendees when there was obviously not enough room for all of them. There are some angry tweets and blog posts online, like this one WeAreDevelopers – A Mess of a Conference. I agree. It was very disappointing to find myself in crowd of people and not being able to switch tracks. Basically, if you decided to switch track you would end up somewhere standing on the side not being able to listed or to see. Second day proved to be much-much better, but who knows, maybe, this is because of some people simply decided that this conference is not worth their time?
In my opinion it was the second day that saved the conference. Better talks were delivered at the second day and more comfortable access to stages, food, and company stands was possible.
In any case, I’m really happy that such a big conference was organized in the city where I live now and that many well-known speakers from well-known companies came to talk. There was a lot to learn. I’m looking forward to #WAD2018.
Again, not to break the tradition, here is my resolution for the year 2017.
This list is rather cryptic and unconventional. As they say, it is not that “SMART” and I agree. Fortunately for me, I’m the boss of my life and this is how I would like to put it for this year.
The road on a picture below appears to be smooth and straight. There are some shadows and lots of light in the end. Also, I don’t know if the road is still there where the light is, do you?
It is quite late into the year 2017, but I decided not to break the tradition and write my yearly report.
When I planned my 2013 I had “the best thing that could happen” listed as I knew my daughter was coming. When I was planning for 2016 I didn’t expect to have a second child.
But here I am with my son just some minutes after he was born on 16th of December 2016.
I’m really glad and looking forward to see him grow and build his own life.
Except of this main and life changing event, few other things happened: we have moved to slightly bigger apartment, and I made some progress in my career (more on that sometime later).
Here is the list of planned things and their completion rates:
This gives me 47% overall. Apparently, I fall into the 92% category of people who fail on their year resolutions.
Nevertheless, I find this exercise of planning for a year to be useful. At least it gives a sense of things that you want to do if otherwise you are too chaotic.
I hope you all have had a good year and will have even better one this year. Happy New Year! (Yes, yes… I know – it’s February outside, but someone had to be the last to wish you “Happy New Year!”)
Time has come to buy a new beefy laptop for my blogging :). This time I bought Dell Precision 7510.
I have a history of buying Dell laptops. You can call me a fan of Dell if you want, but really I just continue buying them because they work and I have never had any issues with them, except of when I spilled tea on my XPS 13 and had to replace keyboard and screen. I’ve made few upgrades to XPS16 (RAM, SSD, battery) and now it is being actively used by my wife for some photo editing and general home use. XPS13 in some aspects is as powerful as XPS16 and at the same time weights only 1.3 Kg. It is really easy to carry everywhere. When I bought it I said that it is “thin as “Mac Air” and powerful as “Mac Pro” but costs less”. Unfortunately over time I could not feel very productive on it. Even though I could do everything I needed, I couldn’t pleasantly run heavy IDE or VMs or play games that required dedicated graphics. It felt like I needed a proper workstation.
Decision making on a new workstation went terribly wrong. I spent around 8 hours comparing options:
I was seriously considering desktop PC instead of laptop, but eventually leaned towards powerful laptops that can be easily docked if needed. I was choosing between different Lenovo and Dell (yeap, no Mac). I stopped on Precision 7510 because it is real working station. It comes with thunderbolt interface, it is highly configurable and it is a brand I used for a long time. Another reason for choosing Dell was pricing. Since I was buying at dell.at as a small business I was able to customize my purchase very granularity: removed unnecessary support and useless stickers, selected Ubuntu OS and cheap delivery – something Lenovo was not offering. As of hardware I have chosen to reasonably max those things that I’m not going to upgrade (CPU, GPU) and leave room for other upgrades (RAM, HDD). I didn’t choose 4K touch monitor, as I don’t think it makes any sense on 15″.
Here below are some specifications for all of my Dell Laptops:
Dell Studio 1535
Dell Studio XPS 1647
Dell XPS 13
Dell Precision 7510
|Intel® Core™ 2 Duo T5850 2.16GHz||Intel® Core™ i7-620M (Prev Gen, 2 Cores, 4 Threads, 4M Cache, up to 3.33GHz)||Intel® Core™ i7-3537U (3rd Gen, 2 Cores, 4 Threads, 4M Cache, up to 3.1GHz)||Intel® Core™ i7-6920HQ (6th Gen, 4 Cores, 8 Threads, 8M Cache, up to 3.80 GHz)|
|LCD (1280×800)||15.6″ FHD Widescreen WLED LCD (1920×1080)||13.3″ Hi-Def (1080p) True Life WLED Display W/1.3MP||15,6” UltraSharp FHD IPS (1920×1080)|
|DVD Super Multi||8X CD/DVD Burner||–||–|
|2GB DDR2-667||8GB Shared Dual Channel DDR3-1333MHz
|8GB Single Channel DDR3-1600MHz||16 GB (2 x 8 GB) DDR4-2667 MHz (two more slots available)|
|320GB 5400RPM||256GB SSD (originally 512GB 7200RPM)||256GB SSD||256GB M.2 PCIe SSD (I added a second 512GB 7200RPM HDD)|
|ATI Mobility Radeon™ HD 3450||ATI Mobility Radeon™ HD 5730 1GB GDDR3||Intel HD Graphics 4000||Nvidia Quadro M2000M 4GB GDDR5|
|High Definition Audio||High Definition Audio 2.0 with SRS Premium Sound||Wave Maxx Audio||Some Audio|
|Dell Wireless 1397 WLAN Mini-Card||Intel® 5300 WLAN Wireless-N (3×3) Mini Card||Killer Wireless-N, 1202 for Video & Voice w/ BT 4.0||Intel® Dual Band Wireless-AC 8260|
|56 WH, 6 cell, LI-ION||85 WH, 6 cell, LI-ION||47 WH, 6 cell, LI-ION||72, 6 cell, LI-ION|
|Bought late 2008, alive and used by Mom for Skype, audio-jack bad, battery dead.||Bought Sep 2010, heavily used, became loud, upgraded with RAM and SSD, battery replaced.||Bought Nov 2013, actively used, no upgrades, screen and keyboard replaced because of tea spill, battery completely healthy.||Bought Dec 2016, using it right now, added second HDD, planning for more RAM when time comes.|
I have ran benchmark software on XPS16, XPS13, and Precision. While 16 and 13 were somewhat comparable, Precision speed rocked. CPU speed was 2X of 3X faster depending on calculation operations (floating, integer). GPU speed was 14X as compared to XPS13 and 3X as compared to XPS16. RAM was 3X of XPS16 and 1.5X of XPS13. SSD write speed was 2X of both.
Lots of numbers, but I can simply feel the difference. It is a pleasure to use a fast machine. Who knows what my fifth column will look like.
Since that time I observed few things:
Mapare still there and work the same. At the same time performance, testability, exception handling, and feature richness got improved significantly. Last one, in my opinion, is not such a good thing as it leads to the next point.
AfterMapor in different kinds of resolvers would simply start containg crazy things. In worst of those cases actual business logic was written in resolvers.
I have always been of an opinion:
Less Code – Less Bugs; Simple Code – Good Code.
Having seen this trend with the library, I would like to suggest simplifying its usage by limiting ourselves. Simply:
ForMembermethod it may be the case for doing it manually (at least for the specific type) – it will be cleaner and less confusing.
Mapper.Initializemethod. If you still want to have at least some abstraction to avoid referencing AutoMapper everywhere make it simple.
Here is how I’m using AutoMapper these days:
Somewhere in CommonAssembly a very-very simple abstraction (optional):
Somewhere in BusinessLogicAssembly and any other where you want to define mappings (can be split in as many profiles as needed):
Somewhere in startup code in BootstrappingAssembly (
And here is the usage:
That’s it. I do not understand why some simple things are made complex.
There is also another advantage of keeping it minimalistic – maintainability. I’m working on a relatively new project that was created from a company’s template, as a result it had older version of AutoMapper abstracted. To upgrade it and keep all old interfaces would mean some work as abstraction used some of the APIs that did change. Instead I threw away all of these abstractions and upgraded the lib. Next time upgrading there simply will be way less code to worry about.
Please let me know if you share the same opinion.
Imagine working on the same code base in two disconnected networks. How would you synchronize your repositories using rudimentary storage device, like a USB-stick?
Undeniably for such a synchronization there could be multiple solutions starting with very primitive manual copying of cloned repositories finishing with some specialized devices and synch processes.
I came up with something intermediate, until the situation with the setup of project changes.
Idea is very simple:
1. USB-sharing device, so that USB-stick can be shared with a press of a button (physical in this case)
2. git bash script that does the following:
3. A task to trigger the synch script when USB-stick with bundle is connected (I do not have this one yet, but it is a next logical step)
If two repositories were available at the same time the same script (with modifications) could be used to synchronize them on schedule or trigger event.
Here is the code of the script:
I also make it available on github under MIT license. Hopefully it comes in handy.
Just recently I joined a team. We write intranet web application. There is nothing too special about it, except that it was designed to be implemented as micro-services and as de-facto at the moment it is a classical single .NET MVC application. This happened for a simple reason: meeting first release deadline.
The design was reflected in how source control was set up: one git repository per each service. Unfortunately this caused a number of required maneuvers to be in synch and to push changes as team was making scattering changes in multiple repositories. This also made it more difficult to consolidate NuGet packages and other dependencies as all of them were in different repositories.
I think that microservices and corresponding hard reflection of their boundaries in form of source code repositories should evolve naturally. Starting with a single repository sounds more reasonable. If you keep the idea of microservices in you head and nicely decouple your code nothing stops you creating new repositories as you service boundaries start to make shape.
Taking this into account we merged repositories into one. There was only question of keeping source code history. Turns out the history can be easily preserved by employing
git subtree command and placing all of the service repositories as subdirectories of a new single repository.
As a result, team is working much more effectively as we do not waste time on routine synch and checking who did what where.
Conclusion: Theoretically micro-services should be implemented in their own repositories. That’s true, but in practice for relatively small and new project, with only one team working on it, single repository wins.